var/home/core/zuul-output/0000755000175000017500000000000015157223453014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157232745015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000241025115157232572020264 0ustar corecorez5ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gf %?,Eڤ펯_ˎ6_o#oVݏKf핷ox[o8W5C% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ_oa0vs68/Jʢ ܚʂ9ss3+aUiE߳Kf^?·0* TQ0Z%bb oHIl.f/M1FJdl!و4Gf#C2lIw]BPIjfkAubTI *JB4?PxQs# `LK3@g(C U {oLtiGgz֝$,z'vǛVB} eRB0R딏]dP>Li.`|!>ڌj+ACl21E^#QDuxGvZ4c$)9ӋrYWoxCNQWs]8M%3KpNGIrND}2SRCK.(^$0^@hH9%!40Jm>*Kdg?y7|&#)3+o,2s%R>!%*XC7Ln* wCƕH#FLzsѹ Xߛk׹1{,wŻ4v+(n^RϚOGO;5p Cj·1z_j( ,"z-Ee}t(QCuˠMkmi+2z5iݸ6C~z+_Ex$\}*9h>t m2m`QɢJ[a|$ᑨj:D+ʎ; 9Gacm_jY-y`)͐o΁GWo(C U ?}aK+d&?>Y;ufʕ"uZ0EyT0: =XVy#iEW&q]#v0nFNV-9JrdK\D2s&[#bE(mV9ىN囋{W5e1߯F1>9r;:J_T{*T\hVQxi0LZD T{ /WHc&)_`i=į`PÝr JovJw`纪}PSSii4wT (Dnm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁eor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/e.# Okx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!fc̖F4BJ2ᮚ苮p(r%Q 6<$(Ӣ(RvA A-^dX? I,($F{ձ7*Oy 6EK( EF #31J IGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC'ǣC~방u)т˰vGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0NH;ݹ/n -3!: _Jq#Bh^4p|-G7|ڸ=Bx)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==ck vz(vb$^Nyo$p[DtUCE9sBz%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$Wv8FHӜ"D$aǽO8'1lfYuB!aMHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I+mI/K(>x=16>6;MvS3_Ue6i'VdW꺺k q+~xGIXFeOzJd z 5զtbJ-42?V5YƏUX<*xGy~Wؔ4ƪ(9\\qv;ZF+(TqV^$~cTn0Dȷ̏V D$Z'e0VXuF*sVBcVf2ȯS@ƪ.3ߏn~]#X ^dzE|I5N!pG)7e?<'a]UAwtJʰ^蔔^؟a^h0{k|i5<_ ؋@ Ko+KsB53csp4 \mqeX0CC{[\jk^cm p< n` Ct[BXkUYQ\B[.pt5,~tW ]._j1nJ 1"c p`>CCTg ^ӎ]ަ%3t = n7Bn>7ߒ~oɿY!ls1>fY-c|j*PÝ[|(Sb~+TY[AD~wϖ*KT&)dH΁wg:?TWuR/uR/UҰ&k[)'E3[uըWX#PVc8 7*O/ōIXi ~|(k֔7J"Rg(D Nȁ)V nETNrVA|b2]FlR-ĤM.&e.y|e?eC]~L'BLUgtgA9XyU*ZbKb+o&aVLXIU ojX:H+ )UU𺐿q ^ IxΚaVO'Dq,D-ALlU %n FT .(G?@/(؆G Džsb:9??g̓W7?<),~v3imo,3h D6=RMU44gIs`:d9%K'(]N Ѳ4U^e3F*zsC܅6^aőGP)T7&h.7nI,wk@Ǣ*#rUy!u \$*_@]Wpx[ 8Asz\uUg؟rKAۯGV9X9c#K*⊴d @p|Lnb/E4M%;g>;k/F%__n@)Oh՞} ^OpERz$ £$R~'QY#/6D_K;ݮAJb̟ #@rz;A0Q:/DY%{*[9|p}$a(aaަ,Fy:H9)4 Qsdumn|yKa)82I^0c{;ypqѨ/ gϧ3s{KHsv&3?OwQ7A F1,BY{=w2]Gn*k}/MŌj&<ϱ~KSBu%2uNu98?:, -u }f+,J#}JvjE@#,*( Y$#xNC.J8۴dLWl\iDH*KI]x1$oB:MۍQ ?ykwH'5imM+ s]υps;Y,SX}*+ඇk!,jm(gɢK 4#z6LͿ&~DQS{$jA k:-Y(4b _4Qi&Y]*xpFqlxx䭲<*7""+1pi5/C&z/݇ToQfہq l4HKySV*LMDjjtQqSנ^ObΎuŁ~x}4VFIw:)g_x:e:`|YqwwåXr2z)r )htMtK};<ymayO[= < ,=q @l/%e(s\xŒH Y.O"A=9],۶5yA܈˘(ۃӤii߃Ȯ뾔Iџ<2]8Q^"fVUl(\_ޕh!p9TsIMCXhp[Rns0]=eʧweW_ FWik`P[]JF0֛{պJ*㿩f+9^ȶ% ǚϧ3IV^F|XQփR&C$=L?:%5eeRX`PyG#Rb08u'oZ[[+ESs5`?,ȕ9Umb0@R ]g^2pCfV詞zzl`0aUУLV쮽JC![kj7+!vfەܒUm(6ܮc7tnD]׬*vE:di|E9Dqjke=nxzo7'yj].HII҈9jʟ55:NObGO%vKYVj3;>~gh?;Di]t+RSp}FKv֑)4K1 -_gg0i ah[b H&"zEVW^RA+bw5bL#>!`8xĝu0eUhrwq8! ުFA$l0fz󽫺6"9+ PH[Hx 0r_DXg]%&1RacN1@=Τ uEmˑ)vX3k:Ux03 lH^몚B$zk4!a.=OeU'#N:X& ϕW$qzp7C+qK*:fSeygۡO!ig+҃+]oȗA-G$|?F`p-iu20@]$7Jj IMU8xa~^"I84}: HP]{b CiW&„rB8İED@z-IY|z] s۶+^K"9vz'LP$eH]Rc={Rî,1{4H;Gz:Ydט}5ļ/K ;7O>_go~:8YՎ,9]S*6 ayzL)9boz bqsp4J^W#jAkV톍hA!(U {VX_Bn {(`XEe BV@+ǿ[?(H! c90u5I @XB0} Ay k[`" lw5V;"wa rD@\!1HD@ #ڝ"2օob!& 7#5+rnD#@(#Axufg%*p l8 W gwD^Y:ph['PXB+ lN??rn J:pG(]MWl ~Q'bdߛ~\\[x;fb<]=v]X{ta)e l5zoӆ&Tw Cpgj"b +aN~VpUi& I8&:x}e;3Q mj22z --9T ^G=z`}DQ=| J2EyH(ϋEJ}3d!9%l\f* X"PYܟCqlQ Fʸ<7t_f_}6Bâ1Sb3hV}Z<iOQMJJa&(0L3<7-.pFbԧ.g$ᆞ:|:48x%`ȇqhߨY:IvGjɕ_%O"gO>}:>;.!lX(Ŗ/zC;zd{fpD\=QZ= ?M*M}x"I ?4{ɐڣ3&اMrl]tgIt\q[_ \5NX_ehz'4G3&QEGFs5.2|һ9g9X߬Թܭaֵ<\XyJH?ne>)\[ .d`OeP`KYDi469e(Ҹ{ ,}GɆ7;2w0)989"C^W05twzrAC2tzۡxz괸E7 VP?ϯSy?'j:\$\ܵz7$M;ΈCFME~8&z=6pØd Myi#s"A2y=^5Rc0v&ǔ9Jf ~"[U7F?C3U} 8GVOJ^M>.bEa6 t֊ 2xA#hm'[Of)z4}w(Ksɨo>F/ɲ?- G'qxm2->XyY(E1 u<V^k}A[דh mZB(,{1u|"_a^am 7e/~$hGTAhÈACgRp"YyH]Q_(&`nh+vYg,L6Z IEO? iЬ0%GQl:+O;40 8?{4 aW$O<=K4t|pd T=ش&ӥ (tŞ (1D%rQL JCk;NYoi^leS,6 4Lc;i=OoqHiUuSRr,&~Ztm >SMXiWByVJ[J&(}tAWnSJwUj[jnUϩ]ٕMZZ[km/3%^BP{{A)β: &LA-uu^Pw7Ag n!,z =SPo AeA-MP[ʖe[ʶ&({l AeA- 4M]ŷ3 "Kx Y"57r9 L|{F[J[e^\t4V=LdUQZ6] <=QdIoa0,Ӳ"Q;~~?Oˬ@򎡃V @UPBSu ^ 85uTW9.XtC4͑j8`x.U}tױ[Z[aOzHl^cI1#gdB>8cFA$?xFhNcϐ~3\J @gq:zp&-%Z\ƃ#A H(QXrJ¥)exki`D?JG| u, R~`R=OPe(/^=@[kdpo 1(=@}̆Sfu] NzCbTe!:4}Vu_KymxPs1@<&.f`'Y~(9)厪-ߧ;ȸmj#0B*ag[ _ J#AOP5y%hGXjLϔbjxxT#@3e,OV }MXt";QfI>d09f5@㗛 .=&0 W%30`=19(2*j3u43. 'S!fҵM[Ŕ2^50yI^r4]7ʞ@ȡlp#*ӫ ٲ|$1pi߸tQvwUBNV3KFa ic YӠHK8V|D6D(L?Ե*qj ؇,4'>RBթ@yՏ; tx+ ]>*hƒ>HuioصE3$G\} u6*rJiX\/CPk!x9o > f!xsD%P::m+ u8w )٢]'a e }ݞ/rfI,qqKo\.,\շX5x7op=Wh"A_KU1pۏj0 Hpa |@7PHޥ3+ rZqr HS"Z,7_?qz^QuU)S_qX)uUQ5˺z6Tbl|õ ,hѲ:g;m](IcemZ_0`Y2E C  ۔[a/j2K2itCnrTnwmqcn'Dk Oy߂6$XtZ!]լWZ*=0KTi\jƓUqiqi')nۨLO rupT8,nTq{/R lԷo.\7 Iyd )6*@6884xwem%9W*ܗ\%(QEʒ$Ewn/!#ĒOG?,o|8r_KFmQ󡛲?wl'Kqg?Sq=ݱd~tlr_]6^N_ɈAoّ_/# *nY78އ|-q߇ϫ)oXqʯ kb}[ZHCANN|f?Gjg޳kS]M JZ\u_Cb ɡ%nf'Y[ _AP-g hU4F&8 dF,?Bb ꅐvG"o7ςb ES #di\i8R,nG+譎J 7Ƶ Xn4Eg׭pX:>‚b se&^$,~٫5/p|B1!سz?ϫ-zg@|s2 nK-wE"Ђ lpfnlIZzcz-,Ź"1P,˫Ȃbtʻ\бhI Ie=|619blX:q,8٢a^dlM'[ʮcp&RD15(c^b%K2҉5G%}͘a4-$XlZ*K'8nCgqrODES Lt ^;q[»L13" 1>;7q-9RG1&w7 *ԼW% ZbCǝ;nVwD?`XRŐZGhQ`L5G*anRn9hb0)R7KoMeY00 ,@Z9^O|Lֿ&6G7XRaI0O䘌賤чdX ^8(AP2 G=p,~Vϭ)jՐK| M-<|Xpssl!SֵzvcsTTS%Y#avrH|Fq4j\bφbr &Wʲ:)/>8kyѢ?٭/K0vZYSnؑnxgҗΜMt|ݒt`Fv4)8s=ob+jFu/ [ OZ$[n^bs6J22al饸LK' KOuU5rGJ4:x^JWQC{ܕY†7sӕPZ *p!j0c3V FbNHtℯ[5,8}@e!7wr`Sts o5-zP,M_-1#r~^~wnu: Q_cF?Nn9p "\3y۝J_TsĐ;}F\Tuڬ&7YesOGkB+xQAR9lͣeiS04-/M Li0dRD N˿guh]CbvMɹ>f(Nv_-7NZi#`*bO-JSl%NLgo/0F6wNB̍;)G]e(h"26}qK"HmNܙ4>s`i?*2%:Ymvc(j 55@+Ul-:5c`ik}șVuhִ53ȥS[1ydlzƜc/Bdž$j` wcEW1к`:ܳ $`Zo`87,%4i:z"b#teU.Fec`?7;sQJ}QEoBL[x*:˙('_-K;P )\83Q7䨳Vа9*ZbaGt`tdL;#s&쮨Et Ɂ`[Hii td!פBx<ҌFMnonnAt% x%pR(j)NoKeqrZdMi&фTD"3z}Fzr(>ytBIּ}a9{J`1]R`(\6V]?:>H4j=9)!7n1*BjIdtCSRj`ф5y%I3!ڥul*FFgU`TOƾe'|3:@ьX Fjur$1/! ;wi3ffs h&S *to̎ARIUH"X:ě ^UނN:cU؝5+ܙA=z,MN)K<=ShqWR$ޕR,1'ansWf.=Rf8n^0_39dM&%&Y+M- B%Q|b78Y"Bd1b aU- f%d\X8f'RC0ca4]ЅRJT<`Tn .^ c]R ^$I<ݮpBűdrf, "VjSp7  Bϕ:%ɺjz+KUSx [%2aN,uQ[L;wʢP99 Gg!m©Zuei{ZQxGCOE "Qsmny.R?w2y!|AL!{eq$|qSzfqIa)'!0;1.Np?hܼT,^q3 k|rxȶ)!+pFkjxzklꋉ< {j^Xp4Wù>yquF*l<Kld}|ecsnA X<4>.HKdh5. ǻc!YWkzqKvṢ]Ydhz!h'7[qe 8@jɦgn56Ȕ?_&iON_K[7NYUcѐbl#z2ZWΕP,zOm nRPـ #2gŒfo\>#lwz],M^zĀ(G47GRQwbۏbTo,'yy~gJO|e|y^q!ۻyhnZD;:NciC#pQz3;zOhPݶg~Ts_o< 95?vAڋb.3H8CH8-9at9s'9\;]| bK5ss8JqYI1j BE jjk+'_4[-(_̕ot2IA՝D4֪p&l喓bR+"FhxHՆ= :;c*Jqj%bLe'N,j\~Θy`,>* trIK>2 _q|G}@fB4'GsaSZBkc[Tޕqdۿ0yA$վ3^IaJL ߭nRlR4)E,q}sڌ`Prâh񄲣o#8>i:RhlI?nGKmgӓQ eZT˜Y`Fp {j { ]C>wb#~gCұxF}6CE?Hp2-k+nǧrcb͋uf&\ 3/$(뺑tGyc{|.sϯޓvp 4x^W0<0@ښ7\2SjImkN/~)ArˏW1&3\wM wa?4v_=wa'35fxQ"di<&!ߜ2@,!;=Y>@ w~hY_%\\R 7/p$ɞ^cϽ^E'YA"\sz5H !.h$UE2O5qڤ;-+m׍KT ͆vȗ^dI0`XL M,(5PgUQMUm$ yJ([Qtz%z~g=dw>77ë~e;BzAp}0] ddW6L. zci "}$na7a%هjjmLJesA'U$ ~:h'!I Oҥ ->z7S'ُ `pS#Mo#]j'w|pwowW?E'! MF$,G1Ъ OxKੇ21z+0_W&-@~Y=]=B4tWʊ_hf`*%Py{'? 7+t_*jc3DAבr7rpMifӽ(zfRSJ¼3z'Fn$ޒ/3 M>{e(W`/8EIc+1G[(Hn;=qFq1R]#dmgH:89_Tf/^\s3줝ٷ#=+7C1Twߍ1˳,U'?j ~k3 ELtɩdk):=\Q S  !S, U|zd }FIB@kCqc .-ڥ^v{\/1S⅒o $ 6ɩB|3(*^Z*PkxI\IRi nbq^9PΕD\E2*ȬUw㐦^()-5UИQ #yFJ '*ߑwY*; ?Ƌ鎟x!|7|o|2U}kƇ?;/xhxW8H+0Z!j% eJ|V?KR}™n*RQQ>Th5ku*WsOq]Y|? j[a"M};koG6M_첻0+yAQ$Z nPL!FдeM׋λo2|nEIk K8A\"yi5hd ZX%t' =-Lvg9jd=ɴDBq ׻uiX5rH0-J5xB( 46rd M3iOm:A/aޙ{+iyqYH-pbrKII/yj$Yp^é$}RNT--(r( B2ISyxk꣚y>PŸO-ɫ}@%k΃_dϧD9T{<:!IaeN+%lc7T2GU7Nwa!vU3Ky4H)ʝ)W=5!_s{Cof4 gOd٤sϙΡͻkvw{Fck6jzp'"mw 0^P!g8OᜯkӺLD gmƛs<8mPļ+xi"HVS\vi+J.;8?t1Ϙjkxb1uΘ<t^KVZ4TKL }]4:8}e0r?-x骯ӊg}%x42l1I$UԖt}qtyRGrSkܜ> xuhB!M.%[XpB 4"ސF}]݈f"[ғTE~ֶNa}p];7h=&VΔAqmR܍zKrĭy_sMe5ü_ kU#[3CcL^ c/=~suk5M֬t:^ŧn|T vV]B#-qX0b!w$kͦGV6pJ)NGE'wC;]OyT5rIT=b $:e=#Ge$ly䊥I]JMDFxR;U]KP/'Y̻~Vͦ[h`!Oנ쳺srBw2pjm[ƼԤn%E5 ɘHU822D|DPIC$@ –ѴFѴ-ݬeVyQ莟̐tՇ,(U@1d!n6Q,-b+mBdyL,0(r38Lizu \'3·qo wz@F*V_ne } 薏4^R\۬v9ܫ_0*}K(p" /rR<~ jO} ]TW2'd9]_nKv\ d*ߞǿF(j F`0̝9 7C6 934 4R7r4VW.̔YK ~IGQ.=䁘b#@~xNוWZ5&PQպWf,mN)xOǐ\]4漌{4ˠ9Ryͣ;KF#NU|B%iqܑ S4ou7;BgO\ﱞPĖ3 .el3HW{㐟o:(|5tU;.v 2Ğ4]Tк%.ˤt].GC입+^U?>'Uk!v9 +![SM֔tAZ-Nt5g] *adm?hZU \f)L/ҝ-ZT\?B$4F"G.PmT, q D '*qM' V-La(~id:\&M/߁U|4S-;n Jh `hMk˦ՆC:"#C4qiDz"p-w[I G(lD\yy޴Oi &P2`64d{XZ0X{XZH* K"kR56-Z* vj&MO9 ISΤ{ɂF SZK6^m _)V?*l%mR\kBiZ C3TMuQ*h@eX{L_/Mpt[btz򰜸Zm7\\FZ:iVITИh3qg1`=ZjFYɘ8o7|1()` j-D0 q.l&D!J&O >dÈ[ m8n(*m>p1HYG#+k5c&>Zn}:x&7<'V!$71z(r#Rx>eΣsʈoEk$-| #-epH˱8#^b8j"Q皡I)"C,"=ek% z Z  7dPHe*%[ea3S$4\7 T:[杳)K85'5 RFG F(igr{@z( K hJM?<Na#31@J TA4] oH+\H0\!lv7@n4xȒF(Ku,%*[W:ۯ!)g5G*]+}$5G6U]'.J4өq,I#9h( D ܀נ,& p" #N xl&FTH XF XE p>A9XF'.H~W?:bPD9#K6h*5P Npf !]#B֋jLhTdxN"Z!Z$YvH\?;ƌR1Fɢ"(e,\aCqX8IIFH13*Iyŧ( ;ݣ ʦaМd9M)QI8RIq&2D 3M33MbrJQD-zT4WBeLF)ņ#Mc"ZVTD+(#3[S1>y*Yn)IFNi5${b俬_vVL/̛[UBZ&`v,N#][kKG~/+k^MӍT4}Sг6v:> 5`:]l#\Mo{;|F}ϩ1ɢg/F}sE5 ef5/z'J`צa}k&}Yۣ;aNkO8v&c:\xkZB%d=>}Z0>=/$."?Ô5§Hb apͧ eQDM^j]*ڭ6̣k! p hjwL /Cz%8 *%C%X[oHɞ6CZf8ʕ'\$'6ɑ'X  zn!VK;ylǩq__ Z2ޞ5n#gr+ =# 6"| xϣgS$}W.|ϛx˯^dk/{fl Zu}c]|x׳Aнx<*M<^|o \t˞Њ/^xJF)>y9L. |׌'S~fԵh{>TCqJ:]͋?F|ax7qÖc➙t.wWE1qw ]C[ƘzyHAC (rcՇ@3U^0c2p,O:_-P*R~<ю) ~t_B}w_͹:^u.LDν߫^h[?̼JԼ:vз.ᤝX?zsI ;Y<;V6~FʁOJbJ9+HУK9)'%ʨ4$]F~W 'P[ijGuxR);6 :qrv'z]9N7̑/A;O"yC z n'b!ZLPogY[&=]JA(̀ޡQ}zuN:^=Wn>F'X{X-)%qgQOoxH߿1ݩST чGj>Yk}w( Õ{~z* 2;nbop.6p @#irC- (uvn5tre-(hzI֥Յ|ݣOu9amϲ2?+?˅:Sq<^p,Jg7l8>1f:8g46,&MYM4OF~< [ C *?a JForYWVVE'ߒℕe/eҬ$3PR ~ 88>&D:tnQ5@7zh^,̄~)S:P؋iF~㜕S|bL\:ۉ:WϯLITʎjXcH/r{Ѻ/%Fhm ^aI?  /'mQi-$;HS:NݷHpJ8ߧ$ʂ)}N#hm>01Gz%UQ%4- a,'3ڡe 3ӳ0|*-l<D&&!"1$H@XNFIBFϐ1I4dFXKSYyԘ X>d7P<(O#Orx:Si&y>:rkk25?v Za(رLք|r 0FOpn|XH\ E%DuԯVPԁ#M;us+l="|ɕ8IWŒμW-WZns*/ ӏ.<0$[W8=Fo v W[=z(Jd88|0E=s,x$R1;@Ł6T Hdp!)~,T7%5 {x0~}TΉӄګFߞm>B@^L{N(X^8Z.+(j}c3EUE/\PZ*P~(Clq\bD Ɜ窢Q=gH5y0>>1.0VyIM%ߐ[@`b>`Z?"Z)? pԬȠqkOׅ֭hkgjAUtς(Mo/2.a1^L 0&C>0>X^. ?5+|J^Ԭ.0͞bR+\L_yyAA@Լ`Jhh} XbAZBv`N ۇ 7(2x;aZP\:D:x'-"ɇqNX2AEkβ |Ok%R4{= P>ގ@u¤~t G {8TfIʃ[~l=Uaa/hr6ԯVP|l 7\11)% 4_MN\R&{!U%=K^2Q]>h三 E/p)܆הKtA9vQJ7vyW_yߣ[Kذx"t[{et8O.F׽'] "lj=b~NFSh3Uo>i0?jH&?4n:2M\vDun?BIv[YZ9yZɡ{*Xg,߳]TMK2Hi7Lr2y,۠s /&F;%VIW$Y 9J UxP2KxvN\ڪSGeXVBΆaQPMTCWC!Gk*qW!gCF@|Sd2UW6f\MKSTh*flh9vi$ʵ$ݣψ3ژ[SWحSU湢Rߛ *yicc6fom^KS4]KSPh U=Fm=kI*$kI*>'RRNdؘ[3W³v U)CQ՟'ۘu4{%"H4%J3)G mQ ?בq%ψq1{j*нD*6=UPUh1{h*bM M%֔6focUb֒jωf1{kkj Z2+߅/TťXuz5#Iχ(*T"kTg@UAp|hԚjTRh*bho摚y#ٓ`#Z7{iE`ې$8:jL>uMCRFQsMRĸJUF@ n|7o 4U4u·klS\M2k(ctSO7ìDMx ZVɱ|WJ ȤޠH321ׅxxs u%m~gԻAٯZlE]ekg0?j㩽?pE"I)WKcc0!(a4Q:eIbPʍ 1H2PyQ`}V#HeNF+1ev`'t#PBN;D"j|rh.}d,yY؛K/Te)TH-JㄥVg*2g 41LPR f,!1 g 4 U@@'0Z*ITuHeӞ%k`>@h#WNYV׶0f@iKhRĥpAJkÛhU5RJ-@rZE.lDພ&\Z]Z2jEto?5SmsfiF֒4E *dK [hMjnq]V5uXܛp ɋeTbYKVR9T(J,;>P9c4HSo0`]Hpe%>ޢfs:] aa󟗇8yF#@e\G"#:ܹ?^T?;N׻7QhP U"ePXJ=vDFcvZ~ G~Jlrd~uގBeS`۫FnճgYapK1ξai3ӪB$J*? V|>XYERX˰On5 dcŠBzڤDGU)yܝNו1tc-Pf N*e%firp(J  Fd2_*y+ZXSZbN^m.:hn*Ҧr#&ϑr#0'-nqx O⼦V!)r#x(EA"K=&.`BX Lk PeTn]H_T_ !xϳ۔y}lq6ؚiEc~&7{Dyʍtpq1v!!5RL٘fe #04iTrxȑ$,JM b D K%-,)ҹg%6qYb$p:@Z[ KBĒx*%fwWBRl\?/)UUN2WY̑P ֶ''D(ȵ 4 ĸ_xY&q{o}]] !-h@hhS%7|U/E i+_ŗk`;U,iM×v! ja"ˊir VȞ&f, ѷ6JHohgX86Du* jLl4JZaLMS&lٍ #2J]1*2[0ŤeE5ϥ'LRd*n~ 6.\oa nK)bxgєI^$_i~`ȃnI'4gK od~8g5V.`ZFH뭙k5)s}ǬT&!RQm; (Jv* RaM,㚪1 ڤUR{'<쁇;߄ ƻ=JB4ʦ@~`Qcp, m@Ejr`_hrMJa; QGD 4,4TUN:p[HnG7F$V@r Zۡui42Ah=ʎh5NJ45(!:Y.}9+( >KXTeG7/tD|H(hhUʢH đ4ɍ@QfhDrcqcvɍ‚6~MNBHD 4:ct%fWd|3E0P=ӭᧂ4)ƁVV!AԲW-)RBաPseCTPDMhdP%PD8#OŹΡ#B I M@$0ƢP((Y렰*M1{_My;KO+eEbD(!Ar' bn(wXwxxӳbr53ăhqfjr*1BzJJ*wʪrd:9{˺.@aA7uB(ke!?plJ&P.PؽVc :Z.ML$6BHo@T|BkiZ"{O%7lcg!BHoeJnte?䑝'Hl7̝21J.7JJ夌=(NTɋu/f)޸= FrZF4Q8vxN)NyIKL”8oӅ"#ZȘHhIEE(t;Jey"Y])2I-i Y2W(2i +r)2)r=7dx7YPq@iPoLR\'l%ξBTH`U.Wؔ>BHoTt*,_RRM;Qa-o琜%1*/ Br>]6)|n>doA*3M&(AFqUH¨TP!Qh +tJ+;Sh[E%FAB B'h<)BNf#it)J@ iNwPQie oJ UF@k!଎P!QeژVJIr߉*[)2M-ZUURq|O /<?MyO24ACO{Os`+]~Jȝ864E!%AgCx ߻IܭmUX`x5N/ϥS#4 GR^:{~տ5?gkoA?Og_|_A%ϻ/|R~k' EGgm"SֶJ#DQg/ S:w9E'~vs2[9(?(A;[_eKFPzC[W؈w8iˣЕ6 .ty?_Ƴ~ZLJ1 )^  Xş~'^ e6ˆp~a_x4wz.Z!|St~=왵,뮞 AX=/_-(4wF՝4#~a?NH]TzgD}ikYѺ8a8چd5DA]J,A 6Z~a4ߍ(, d0&qo++lW'e0Vꣾjk]nJc{-T]Q[2ɨ=Wj<$߇%/.М-Wnԯ0yg`YNw͟gQU }:NTxrh]f5U^ZJ}cJSc'} uսr?{Wȍ1`l⛁{ !kFN' _[ےe5eyZ4Ld-=,"Us"LϿc Q7 ,(L ͽ Ez"S*)(b֎H"\ircբK CVZk \ hRXmٞcXVjeSj2@ƑO72wM]ߝ~tSumP%Ӧ>IH)^HDf/JTG/sFj+ϴP/c/O&["h_ nvкԎr|0tNh޿ܐiLv]C&웘^^?dL|Lc v Viq%0#z ~l#}]O=&~Dg^)zK_ҁq_Dr}O Z6O;\=+ђ>Ƈ0!Wm}cQ{n{LƔls⋽_bR7WH@CcWt HɣwFF|v*p`BW )J#p*dM+o@, ?̹b۞hm)b,K q a?r>hqAacvwy馟9DcʖYDH"uϻBdj`*jpcUmXi0]/f_:4~CAV0 R~ UX-V%`3Ūk69?| Fd_J|>sBe Y#^g?miS˶ۥ"U¨',Оx9-d{h#c,Lj'<<KG 9t,{ђJ崐: R)> ^Wg,=Dcו )ƤCs )P3h߄?\^uP<s6P%n9A7&jPk@&68̙x(rR{|p2?z=4*Eƚ yG_&77J"moH)d.sڽǚon?_s>:gl`/]H翢Ou W?ƹ,*o+Kڳ;t20M鞅(P̝GپT5XiwƖd%Y1$˖d~;ڻVt..tqe_v]\pNGTvhD..XťƊse2vqt%)c)4 훱\=){1ZA6v9 %j$hep1% nm`ɡ+`RP7 >upq 8YOAM\j<9Ićp/ȞweSbKyZB%V#1Q 9 Ө<= eUp5=(ӨDw^rp/x\fÿATʘ'.VE [܂{Ye,YYsjljra?="IqW$E9Z|,({ފMU$)F=J$1!Kv]QtlDbrEr/)2+Y`G;vM5‡^  V,z4d(N {w_֫ncX5; ?yw%[ $kidh7"I+)5JLXiDeKLp }4+a`ѬMk<(D+%ۇ}ϣ>_Am:eSǴ=M|~RXR[햮uZX}lڗIcOa#ˑ=%&Of}Y6/Ɂya`b`^XWayaZRr o`HP3ϖ֞-Ua~挨}Y:]qV*0h`)`M+հFDĜq/k%LO5_I9IXLwIɚ4l/s gnSŒzB9ݛ_zB{]Qb%K1bLn D+k9,kpPb_֢x砜/Az,vI4AhxqJ#mauZ\yCNiM~.̦xe)_ȎMq{h hRW@{)_JJ3ޅvWm 1Ri0;gHZ?;f< fgτj ŋU:^F08 8N׈HAu{v&+NI> 京#H]eG ~Kc鹱 $1r:t*&x{S@U0 ZX 3pUzܐY5` 6*QGO_͕(uTV~Gv9ي\(ku" R复F}I2Ibv:hz'zWe'ADc:KףY.7c \͇5=OZ^)m L"zplJ[(-ڰ~pl sh~ui8]|Z *1wGMٻh#/*C=n.4[L >Ćyչ<><;({]#Pc%;$C5!!w׷k?_vh̯)&On;F7y8v-n1~bnvh8fMYzQѿm泟vڴ y!撤(\{w:G퇟9_=~)nz5)u}wY$'?wnKX>Nm8}= WSzuOOH>]MP$c6ו" }Hj[}+Eyu/s钔SBڇyhVc{Wu H3͛]\\ > "-JAy>|MdkJ/-J7Ox_tӘśrdD(Һ(ry<[e:O!n!yrdb=ڵM~Lsܧr9MMqLOG}51 &zz<{c^n['^y I -gQZ#LVoז{L,ZD.\d۸R;#Q,5ʅm:aP{Z۩hBQ")4܂Z4 8 #ZA.{g+ZD[2*r"JNr2DυY21,A)<1)4Z]b,[KkTKz&QIZ4s1 ʒlCA+c|H4Ѵp1gsY;Ԫٴ 3rQS.CjH @ I,ni/t{sBzEI,tU*b%}!#Cc$r/ d$6̄UK ;qhVC(fEs5I>`0DvY$0EZB`>]`[U c3#)0&4 `wt)BSQ G eCX>5z`3ɉa52dԕrI2 -IU0!kRDKƊ1@h'49 Xb&H/d#g,C҈2zl$&D v(7U#FqIG'bP@[iJ lɲd%$-&Cj3ʘ  {[ Qе`i͒Q;]`$&Z;D%Ġ P`HؙDQiGQDR w2y&a3(/D*JDM0%c(őeNN|ZՐ*hܩdi2ޖlD3/ 렼 )z H5'W53kMxӐ) di!ʞ<'$ IENi :tZ?{Frl !`5l/,iH5Ԑdɔ j D9N@ e0>A2a+"Ǜ1sՕb1NԙXq4Ɗ~24cVGr|T1|,pb: >L lcLp5Y7JF=FղB$qpJ~$$t> ^x1u`KJ{2ӬdpҫhAR0 h%40(yA3 v;XmQ|E (%*x$L䴬yE(C`rQR0vʢhxq `a((AJv+7Vdnh*^z,:W& :U+ϓW14+JXmA5SRJRp`Iݏ/Yzw9TX] XXkxhHc#. m҂^A ._GW* Feu ((5&Ӗ먊{%tMކVhGc3y2Az5|< lwXEyWΨFZNmJ D%ͱK,5H8K&JI>2~ЃA2CƠ"RFdC0",0 E*C A1֕g[~ Lr!ݫeaB&0QIP@ěWT*-=+V^!-IρA@ fM:)K ` \:E?HczfdsC <%\rY.1eBAt %0'B*|Dz;`մ^ÚIQ!|^MW`E^H"\x>WP' WBNa|;I^g<[0jaB'BPYPcx*QGQ1⓪``Y:P}Eq܏ "e~8QcsJ7H0r#Sۙ=Ԩ |,&˃ >T WMIAp-c|er5Wἠ)@tBe_j%ֺ2 :rfF8Jf:tzAs!=?HoQ%q(%b D2тhQXW;o\nvYJWsp)L&%fO[[m4vEY~:A$kll7/S. ?~AO;D gWל-rg[ڻN$Z \4Emt!ɍ@ p|rŽhqe{˚w$7Ҹɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍܨɍȍ=ZzABPpPh~BlBߧPkD4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4N4K> #!ECh}BBiU:&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB#Թ]*w+5:iJ+Ar{z7FTJ^t,J=+ҝ^cP}YveutW糋dLs kݥ[Y5p(`i%/컻`78A򌿌9e/0?ŋK?v,4RA@b+Ȩ,'Qm9'8>iũ(X @-*x>;:BXx@Ja-j(nf(n`$S,H5}wA0C9ItN^Ң΃=$%ҋX!tC+L7ZZ@i~}weWPf?/X9u?_ebq~)swI]`z'v{$ގp;$Nh7ELмox/oNHJus5f찳7'Xq?c\ .@ E T±U `(L+P|(کl `2Ju*7Uϲ'X %Ky9B聀3>*ZVkP <@Zq4m%:^km TOiD?X @ !fVj/a(xa-H X0a<kTy.~ZZ̦xo+#ohmIw[|b>m]gV/u~] hgw\OopQ.)]ͧhFo{GWU9@!ON2uYy_7'|IScnj}MwytUܻ(9,_u;l#K\=&Y K*n.61v9]\q09^l. hcƿ]6H҆m+\4^~rwN /s΍1+4X7?}n=6vtѪ5xSi4o>_Ȫ7+ׅ͛(c\ ܠ|0g'e8|4~uzԎr9i;3#y x7C/1r?'-rե ЮnAs:yDO£4<] $: 7҉[VxOI}i.nN FcW*F ʲ胖"['&jk/Y{x >w9Z/b"Yr$% IEhb"e)eSn`VUiRʖ2)sX{M mZlhw ý\J(*PyD})$cN ϕ*\e< ^Gcz9i^]9..T|VSB[5YELJ.\Xc%!l43 a8df؜: h~"iQ aoenvۋqsuƬCq|U1zH.+pB9a9}Flۙx7uYşVlSĨ7J.db,)!ka$ó,yi`K#SV: .z f 5+II6S+ Ѹyd eKineAq *auI^z ,TɂNe~T}2y";*ƝbA@f#AHI`1p_mV}2,d+cCXXNz+PHy⦋@U# .͹*$P` ʠ *JH9GL FcӿOk{B]GUF(Vo%zsQ7h@Ƹ3Z 吾4/\@aPB^f (TɻP͡"xp9\AD9D>+ wQzAAJ$ IO p+LhPX7nիbXlBVH XQW2Υ :9 P 9 %I^k"<2ahaK>'BTY1@J tT#.R1CzF#ׁ ~ xLڰ b"GDzS4Zʈgj Y*L>Z.r}f $@&]P PQj~BgAْN)T_%K$zC Y+(m@~u#F`)1-,02i|P/C>bNčr^q|T)k2AzJ%bY wҭ_wD%rU(OZW!b!m 104+&1MP$9D K]q)z@BF1r^=BWS}"!w6i 0P5a6KBeq7,R~mYA莌]=R (/9qz/;GK":".]0B" ו_Cˏ5X/;[ؓg]g&oVu*j|{pCz]O WzZ,_^\^św- +9?Y!F͗7 :lt:d:yPfKu07N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N3|Fd5K6Qh:@T3:F^5N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N34N3|F_Ʀd`#c\G1:ul0jFQ2fjFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFfiFff:|wFeįg/~ו}?1\/70NnX/Կ̫~U>o|Wng:.HYyJV"0Tr& ZL,6T2Nb33JLeQ2eXW`;˚uTzVBN(=ę=q=T@iXÅv"`-DSI]8`@z 6+D ʼn5`%פvNXͫj"` Mgm]&q'N\n`9wl"`&VX)SYRx2pDjv0FЩSY'<,"P:3r+'`œzp$f"`VMeiᘛ3A*LL[7Nr:1QUY͉KX!QDJYwXTblv:1>`4^s8X+¦NZsb"ϫSK؉)+q6ʼnuHk"`O=?{8XM!)M%TGfs XsHgVm{͇ f|yqj2cs|z,7o7~3Z"S4 츧<s'= #L!5E&V5D)9`n&*Npzl`kϙ$Dr%ɹ X*غ3`RZʉu^MQuuʝz{Hw^|"`tb>{U݇m.W QjyyvyS7KǷ?_e權rqq5Ǜzj\F)@ \?p1RipRihvu^^e(]b-JCJt}[g_~=/mw1^Ϊ@T")]򉽸N#aԭnofWīg, i6W̟_ ./Wůg;j'Wxߧ Ek2y\޴Og9bݰ iЮ zzp|Z@կP.`kDE뫛v 6ej[tVn[/? $/c]YmS猢.Z+ʮ'9(xΜ˩A}Ǎjus 뤔#t|*]=חnGX7L"#{|/0iy꾇 H$B*jdDѷM*d3vS(-Up(ufd;(V9v>*@J#CߧLkcSDS0m}ߵo֓7>=_\rX0*jn*YD;>71Y,ťy<:}7hXXA[js&!%u+rMg5ⓕS2|gczuz*_7s}-xHTEDrYO07Ðﬞz4(?/t\zנ([}ynQF|Q??R#iP:(:VLv\c:^r J-$-^AP5ş/<Ԕw]>l0{c$=Bg=;UcQhG㓃0 ~k\}-6=baðЖ'=~10>_[6eN==B[3x;%X)G4C@;nqc~jaW3l񮮢&zpšͻջ٠V(N5ό9@OGR\ґ׈E+v*$ݑz#dlRGaſmޜ 9zsUnjoTߋr 'JcťgS霷#b .hRdU6ȓs+_y|Z_)^lLwUO Z/~O{\˧q ǜ?ãocl=s?۠OUkq[GŘ`-Ŝy؝ xSc;/)ɲ|]nYAwTWb[_řˋTmʫɧe^%W˭#"/@* |0͠zv`'ybó[u_YkD4q+{P>Ż1T)]ƬW:!HZ{)Oc.Wknl>Eg~NEUuX>Mn? q SY\X_١//ŧysSB,m4Dk6V΄wErF٪T֗yɕ7[#^ )?h{/6UhmNOffQ+X(j}0^lؽ,WUB6Me*)5k[]zI +b aS`0SFK)@oSvw9P 2qUwnfedݤf)USQOVˆŻOqDFջGLCwZ2[pZgǵ|vϓ]"-j[`RPq6T{AΈ rؾ VΥ0ϋi4_/`‹k{2F׮aV] q*>Z5 i,;hRzG<o>Q"E"Oy3sH+#l :UYM-mZ.=E>)t֍ j׆eo9$ANq7I;X_}H:5`5\(wp)De3Ny49 NVծ[;qjk`I;oKyls!%:t3}v ZD#1V_QJX^fʾZfNP^q󻩸IeJuudY%Rk#O8b cM\rkx5ivzj09;>^=r?ԝ~c^wsunK9ō8`1waȱ%{gѢy2% 3Bs/h~|c,E|?\?=obwyѝOsenp|{J=5*1š<`0MT;3~ Oιy “mBo{,*Yo)C . 2Xk8GAbpR;coCդa@,ßK*LjdQe>S[+ɭrbddHSOc/~/EhG?h`ߡcB)7j|sny/3,r Ƹ#kˉ2)dFsr 0x oIJ{CկluǃWqn#pc>i:ka;tGňpˍG;2Rwd;@c:wiGqmSfgᇞP-=[4H6_Kߛn&ɓ Ca]JY*ӻ.qY08U9MԪ`Ǚrr\5Xx\*5ԉk;z+$׸E9s|e(juqOy8Bl;eS 2-U+/ģNÞ)0}Xz60_w?CFqyaCћgF%۰k4"ged{JpnPJ;ͷռ?<.6(f1̢}4<7NXg2ᆐa݄,ƽ*>v.\)tpHj A"Ksppq\QKCx4D#*EZt9Z: ΌL;s'" Dž Lc l a6q֟nΣ癙lMx$`!ōA4­D&#R8f ##:"m^t==KC+,.'xɀ; P s z##Th=1'q+yȻJijV$6F-r_a= ( !CF&gJ:53x#_ȳ ;ʗp]`LSzHHsF EEL葻(HWz}E0%0/o(;2|s0K~^aYvOCԾPFۇɏ~_NuaRClMJv=5Hi\y&tR2]y{={G-mM\C!0v$lTiFiu(dU#622WGC+皡yG*֟h˗3C74τL"oS:aY5qRkDFbw`}ƺO!c0B*@1UqNu5PDtfB4b\ kM"&leNٺ<~ĮfrU*X'n% 2Kos!6;yWb؟g)Q2$8{wܢ6D(f"q&n%80b!:&]vA\Z­GyWb޷oTM+>1O=4`i !>7O, riI{;=j,!mc "nԷׁ8BX%u5C7n`Seq`YaW{cat8{{:/\ 8bݸQ&@xb'" ᾍAG Pf4`(k]j ! nÏo]-:lT{ bÜҤ۳h! D|NRн{uou56}j7hIo &KOq-$R ^i AqۘjJSéۋմ\pho{e7n: Z37 Ar;T:|cH-n )D~ M̆ o¯7MuNj2R८ޜ=_!.ue7nI ^\^ƻހ@{[!.qu^V9yF4w{B|Mb1TfC%R\|E Oޕ47+_)a!:Fu!^p/QJJm9f1yR%w\m̀:8'ZAOX8(I͍3qAhǪ|bW?`x;ZWG!k CUu""CF4obx-Cqo"|/AXS׃zWǮԒRN$YKRmԼVGHAh)(򑟆\GQNqre +?=kTje/H=RrBW7:y|鯵xOrN6; D7)G.WIJpؤևI2fV؋@AP "](!3}u>JWE=-ԄqLᮓm lnRk̈́Jb6♶+"y8^^Pcjru8%V^;4EfI51O(j_@%@ٙݨ%:_dRnD[P#\Z1n%'>,%#(UcoG>V<W+cwGZpUble5nQ+݄КʼnTܠ& n3G={`<ވqj=dB+T^V/$"6ZxM JX7˿ ̈sݲ>]*)CcknQ\5Uc+R ٢G9^jIP&ژֆؠك NtTeKɨ5n"Ǘ~y"}|^-ͳ1y!C|}A$:lz@o=$ \G(va{a%7Fq7fMq!SLRF)*ЛjY^YXEr0Y?L q߂d#NTvYYj6?;YFu+,[g:g+{y^:#'D,^JfOV79y~[|V/.Y4go9oor_EphZ ?WC-F͝*8w^yC>.#gsKyZp'@=2p'mX|}~ū\PMZ`FqG| (`f.;y2*hG>RθPQIe]];Ԕ*6FQcU;2vx[[;vN B8Y~^#O-`4fv^Mܼp^zd8nʳCEvHձNLSsmJcmiG>VS9_>4l?fy`67Օ449x}EǪyB'nP`/bȃ5mJ?à2%"ʃ5qU)[^Ce:cCTЌ^F`ElT>. " 8²ɚ1MM.jjQc-ےf:F*%,zXT8/=mŹ Hzdx!xxǪ Akfxh>~* 4E@YcfQdu>Қ@qC>V,FDlU"/IՅR1aG*P]*xpCUΗ{G=2J>,϶Maܬ7=dpA78eL%_05=2 qn©lIr5i&\Tɉ7aix r*@Qt4 JxeHF<6֘D\9f|9!lǜPc;aiR!.iϳ:$E7ͭ >`j)De)LZ+}Ǫ|±_`qTAr @f¬bFl.?CkJbhiuJ0TҖ*b\p?|9QLjK԰8ymٳHR_Rrs)UI4$HspYz|8(E>RpgSj4fsy(TJTUMve3yC,s*t`lgPcu _.;>"KhJ?V8eN uf:zRvs\R?zQo)<.O;?*;c\n׈|?OCz1m=2m9oFlXů%k*7 `հ\W.pp*|$(H #zjQnOb\Ȳ_Ryϸ7ohAraZ{s?:Tzw && oF":a <]r ޙdړT/sBVQvE&đ|ރ(h. T0mj4kEsi&\|ՄCwmr>fCrjzs2%UqJFݰ}J.fBȃa_9=e“@'RnVVEhGbQ#mtbTŴ4<3,"!#`S:Z u2"j<[w$)V&KaO] Wُo^g ͋߁Vir`9tsTS8:/zjd 8~Z/Q (q-p&T1+ b-:ad;pklPRNn&%8K1r fbMEQŧ MKs2E0#83 E\SťWoy@(4PqBqQÁzd;#TBl&v`IM+tMy5tE>R͖P ,F(@ s ) tk "3 ŷ\h݄CϿk}zFC FQrM^·B9 v:0+?T(s JU&Kw -]@ 'z * iT+ E1@2I3a\&Ĉumj^< FUqbccU)AUpةK 'k^yfct;"f JˇY*7V惈Q5}ͨ?7 *OkARb " O(#i ,*LGc')o X 6WkEEJZsGM?s]I#TE$is=G<" ]η3ZZ#]q} vK\:v<9'q qc5 &ab0j~nr*!E~*G!( 6e1v@RU#0!'l %51N3WxyQ#D>lajpzdܶ^ާ|s.GFIU&48Ghl("SIUys9P;"TMeA^^k5SN*@,EX iq(ze@fn .T 5Q,Rd^a zamjh~[e7|b87g~2[d? b@D+,7}p&˚)be ӆ\tX(.).~'hץw`FNuę\\͟x\VR$M!y8pPn࿒S6on,,fպ.s:D踕Eөx]6h}Myˬ`+aH|1%ѐ _6 6R8g~>Y|Km.\wl XϤZs+Ws(o[Ľ|=̰lr|u bU)hQ(8aTOdQJKd-%٬^d2FVs~[;2[ϯ)}e̶<_>|B|땐&2N)^qEҦÿ~j(zOQ#("<|ų):\>-ǩ.*5iT̉ϗoz3 E6廿'g~X]AΫr βc{/D>"Pٿ&~2{Y pL؊$\bu%DG^{ُ"׋N~Ƶ~iC#$u>?>;nLnl=),={|b4tIu2,oPM#w rғ]omjQd"G)Nl9Ear3N\˚km_].5XξjԢ<#ɚȊ:Tp抇J`%mgY- mql#I?_ir/^9r=5ר`+ i&DZXqX%g3|o' %t^\t4{1Ʉy_:0?g~/+ ihͣgR]vmrpk|+#=2k5{LO6`B=Ot4g-|m 'Wan G.ҭ3^TSZdaEYg#F2M 5i@( *>f|AICy(@+!B;sx#Ax⹜jsevo!c1Đ86rȰzp3W^0"'{W:Z>oP@vbV1,W*5NKhcMݴy)?~Q.S)40Gsv0G>T>gǩwnYAdH-n^YbVuQ_Tbqm5g`/ZqE>Ƞʛ?\䴀Ϧ_,F#}dmٌ܍/Punz, PկwE4µI,|5x!>z; OKf1^~/Zf6ua?[vf'A13D&ex&'vn~`F V`eKaT,iJG3pcO';ߧU.݁/ՃX_AoeaM<iy%Fi1THP1 %CdD;ld^Ʃ"j9 V@`|kk#6ԓv:iLkDĮsnhq͛}TOxE7`a-M/Xpc=!uԭE7M7˵Ծɣ}lsדqȗb6݇oN_ɥY2;&sogof': `Ü-(q9ccfc\ZbS h$)̙.xY)3x$ȗ+_ALᅯiaDKB%!-{5/ᅢѾ4hSTjØb>穁u=vl[iZf_K31XƔӅֽl'pֵCFY qU|tI_yC?270WE94 ʻ1Uj^RE[ Pc=PFx<#eILĨpHd4їO0P ]D/.Aĉn3=w>#9vJLFb%ԉvVUXN8˧ۗXcث]/Y+.0'Կa%}hCi[Lx]fe T/pD !h{?`x>Kח!y4 ~8$BIMݧT"qQ^J鐛A4hkDR|0A(ѻkQ(( ˂|H=#!Ԫ {b*<1cYL@nǵCw-=L(v6JpJ F*? $'Sȁܢj.L P +1XQ?U֊P"e2!(tbH(awhA=8NLzI$40TDĬLBZXl%PCg9L'ճIjQ`A<x~4{#(F$q8cQC4fhmSAP;Z?]IQԇs.;r0) X?&ûENaE 6,CTKE[D_/ipwʆ;{"ʜed9Y\ԶdH5c ;l5AK܃_V[q ڀ6-)bdA6JQzՉ=S^k!I`Th+GHĐ5 UP0ۻ@wS0Gs36˱\6|o6Ѓ[3qr,T T#^Z wD2(}E>fs^r4_,\68 G6C)GsS/Oxd830w3Oepɩby!VՙYYkBs,ޠ$8D(OTQF3%\ayB- ՝6h.,lS%gGU._գ4΢G52څ* 9Zfr980o#+|ϘZ_ 1EJ!*0!BićjOj 3u,Oi#ʸqT*Fj eDebDΥ 最}nS^}xRC2uT9I=sRP9Z:c)n"V!MG KVX9cwI8905`i V:(` Kp#gJwFsZʤAvg*3DbBYI%G!2JZVE yɍqWy#\b@0[S0k{>%zf-d>v>5K8aŀx%`cZTPdxY!NjhgxٝPe'6?dxDUs zu;5/i%cJe]^6\eЄ iǰIHeÄZ?_s[ß}QDS l3(r%*&J"~=^surC Ųei1K&{0Q*pJJG#0^o(n?! './-Ʒs_]Ыb;he}w﹑qq$Voo ';K. zpCe&m)k\wRd >e]ۻe`"Y HLV*}w3#Y!A/fp~BY» J1qΐ""a%>l<姈Ы@Fu}эXu{R/]+O>לepЭë#RQ;#BrQ#ݮϲ2F`Dis !ƾ%vmC5v6i\9bs2ŹCMH>:95t{X9Y`6yw3+wO {Wa"p{7fY}A ^C=ی)=Ar|TG ΫH0!D2U!"iD3r|iq)۰]Nԣ*P#j4Q(Y;Bwx&%D?a)ؚxK0k)jEjZG˿-W,&i-]NK/>DWuDuGe$ɪsSETP  XA(MX{ m2XMGRư/a}sNE)k.q8F1r#2(F>O%Qp0[()X+Fh4UfljCdvZ+nB o A!JAPJnG\1_fi%!k{>]r@1j*N{֝>$'&nӔ{(ߜ=8NL#x4)YDx\#fd:x8I$4<=9(> 'G_8T=tsX+puJex(bVE0ݮ9qbsW1r-, j Iۅ2c%,d'&33M\&ǚ[[pde5Fv<>xiɒ%Q4r<3wDip8S)%)@ -q8`".g_d$2Y`XH=CYwŒ/*|JBgY{ȍ+A; nwAv֋d(HZ5=FG`t*VXRQ@V t*8 {~*q\HDu l5łePA2tYJ1pnaLBN!*Awcd$ 5јVg{)AA^' !@L L>c"QBQ~ʞAuiGUvxC}!ođ<᥄lġ7̃tJd&vTxsi<dz]b:2ǑOQAK֙I.%@_3E*0El2nđ?}ap9,O(Q|öL5Q=<ؠ6?2]!2zT(|gVLQjAe\1xFl}#Ӈ@ll=!`ܣxdPDkr.;~hl}[w FZ7|QM[N_8xi\dƼRm[H|}^m3Q[*q7k] h)82\r'(Ͽu7͖3KVLʭi^_VeU^CSz)(]UEmKsێws0 jrdX z&G,5F.berS[܏+;y?̭{e^[TAfv`q,?x5xm Հ^3< x6rs?h4woGwx}v'G{]B矯R/\ro;4XMF`&Ʈ[Of^];%磉هdH4IɞY4٬{:RyS ԍR{XApuuW&=q^]jp>s9D" CnJnssZ% L ukE5(.r»祖ymN3#_c{AKn_l-jP  W=&$TAmVUϐ'AUUrf'`o!аc0d)˖*f9_ݏϷEU^e:_`"KL%T5ԣ58{7#7G^ǎ!R<k I}gwebaI B3*%j3J*&==lN _uT҂0DPpsfQV`-" ”cy~_w/HŠ mraf*l)kOe!!É#pkAmYduBCHZ嫌}/HƔq@uX\`j#F2ǂ:i f78m Z~UҎ)ih ;P`]c5%!agij7YO]>-fpH } $|Rx[I1w`j|\~;Ooܘrc'~v5,g+L%k%R@^bYXIifc3e_wg ¢u. (H֐ޛ_v: ړ9z~ qnzG/ ZɊVV?:3b4}s./}E {xwx,tw0??RXݳer-ǧW>AQmjeK 3I5K;%n勼: /^4:0޲;h2wVݦ }|bJ?j"wdyû9j{Q뗕Yԃ_7oL2} \'z}\U}8,^7e10];TB1r9 E|ꠑ}f5g5ohP8AR[6eymSd?sTI2t*L W1]/L7Pgm_!5'ȐUH<\!N(/JT i8hDomZ7if ڳn0ylK*:Qx+t! !3oS삊+KBnAk= K4aa&faqesC6 c]&(BYmr_ G3yN 3ԝTaY JV̞,ҦYת ggT8_a;6_a` gfWt+`7O¼QW0N[SURYjRB w=t1= b~ߣqBVVEݿ2Id,";tRj4ӹXܯp)r]* OG0 V5tȐ:F*S"hSA5?۷ נ! xh{wtT g>Lei9E\RUDU60G j( :/0!M F8Pϸq̗3[mKQ9J渗f^^'܏\<_^4enF`̝}ևfڿo.NJ|WOL}m17-h+k_yWiJ̭3RρoVstd9tϲ>;qޢY}N/V%(C5+8LҶH RJDgj!*n0eW/"F_&`8髯P@ J`7?+|I9U IIMͳu@# ![.T\,>y%SDmw+&xA2"6| *yS53T*fʒhŏB CϸT3mb:v9C1OVB?F{<|FhEkbRQLevS`=<2㓲}+V?³S`BGzvjX.搞݂؇gyvreq@fuN"&ӂ8)),5_Zv0f6k!)9dI|9y':yONޖE׬ ʝ_f^aPR3V-eYvh2&3'(&aW/ԘOi&sfD!ߔ*J,O^Ei'938B2Em)MG˯ Ep!/-Rrsl`R%#Io aSf4O%}ƜE IV0b<Gf 1 swmQifF\_isܠohѬݿw'` $l=Cζ~]=:8V*O1T"ÔH~Q;&_ TwKQGd``kYY]=ۺ %i^&//3]vwuctWj7_l<$޶_.pwcwvsW]CarYvvjvnmtv] z?Aפip~'7@aW_lJ>۽ܛJM7]E0*oj,䧍 þdֹq^%R +QVWr7opUV]7ƯKk^ puuWS3K%Ja]jk0Rbl˭TH95.h#/+e9ZL!UYuD2ᄃBշ: @'xdg,P[Ca*h?ex72K'TGKS'Jn$)-zX'' D d2d}͛MO`7h=XO/zR7X]yWYAOloE _;\Z Wz~5_NfqjT`ж_N؏d\/[|1w?_?ߎ>ݨF`Azx}Uof̼{z~D86Dsg\cӔML*|.7:ї.@P?vQ!׀ t[&^ ѧ%F ?|8),y z%enE2wgBŪ<ԂKYT20X2A0yn9d>M1C1Nόw1jwF܈)؉Xb+sqd$Y.RƘc7}x-0_'84lSq L3:uiB8$t#tf4N!$Gi4,󼕋7kz4{!Tj׬'~j f4NTVR0L*>)A9-Lƅ"NyT0ӱ5(]ahȌCo#*%" 1bW:yP[YhB@)=X+S:@޳$+_n},l!/Il}Y6EcVj^!9${3lfUͩi:>qz_~\=13\^m8:ddbHZf&)Xu >4{p.GW&jYs*ݢӃo7ƚYgt<]Lr[L;bF#e,R9õw,y͘mWy߶<XA%5x Y;ؾD7wmQAPE/JߪGI)ۥG<Qme ֙q;=&i~6 N5Lؾ͛0[x rJ}=68+ ؐm!n,EtXcWU=X?M9מ<11UxB|ߘ wKds=E"U[Fn/[Ǎukvkhplog<[i2Or/33߮9!:Ww<۰*rnǦVWM.(kI#MLsP]46n2`0k5l5{4{jT$l FF JԷS?IRz$E WD +S#Kdmg.uR_o_; Qk5с/H^$^^u "̑0oRN W*.*hXPq*+ZN,`lMG\- Ee oG@&`}EUMʉ>b 8p*Q),N$ViR~Dܸ'm!tT8}n?o?l7-+ힲ3뫠ЏDꏄʏZ=N-N!//$MˤH?fFDfzƎ 9F5\Ad-l8e B.+(vٵy>Tނ4n]MsGz$c&ߘy䍩ў'v}|-4c1y7<XN^]Xk¿mw< >8tDU;f+ƣ+'kr㳰R"yKHbpiߵ8.s"mXc%`YU]`BtX_r i 3cc.YJr.<Ş8b־)CTUv/h7~vJ 7Wj^){cKF?Ѝ}×gnmg^vU`\*z46ܭ 7iaC^&J8Yn9Jpa@49\޻ƱpVͽx[Wڬ}j9=TC2=:w4x?b5ь *t@94ȣ<(|CM0<+ljh=r+яn\Ƃƕ~YaځZ00G H(21D>B etOmXvT,-Ћ~TMdS.b'z e[tPF99㝲J7*+wJ&Be3 &V"DɠdY5K4H)m"v+/1Z-26^ g:Ou+cr50GsE|l 흏˗_hp,mLdi67Lk6G'3onٴ@ *F&2AS-k `2(- Bj\"f0fXPKZi YM;:'M+Q3dE>:'HnJUXr,XgUvYa7͐b9 d};2̒j+^[?`I@\*Z$4NWoY޹0 ȸ6HV ,ɧ~VO ͵J!ptEci׃d/9? V[}ZʅOV1Tx~nfu`#|L&aEh/$Z],pfրoȆkCGðg0.x/Vw:*FX;j1 s[ <Zf+g TrT6E[?ipl0ci5W;?~ rZai kS{g=KveW[Xlge$_#Up||F mW7DO}}XɃ;Mc6S=}C!T9/;s u7lN2ω7~nIf?A:,y{>iFn"y +K7lxReVLk~%IUH]Qq$jw3dN{"C*SةA><9? `hH.>`'Rs6U޳ie.8h0(QFcJS?:flF)&%q %ȱv~T}?iS+qp5ݗ_!>[!7äԸϱí33h)Z6@gb\4 [s 1*+ -MbVFoe $e}tdպGB_S/ +d=g?~zLeC #F< T#UTiJD\ڬg&־a󩹡oi۶oH%"un8݊KU*/ )Oz%&K.㥣$7w2sp]`~Gf ,gy83LT/݃3TT$YY+Z%?Q>iX$AW"AD2[Gaڗ%Qϼ31"X[ęRQќ9]W7eeLi"z|`k[*KЎ0]:/9n̛٠\~@y Dh5uG2`w6r*p&Z^i=1"]Kжz? h`< M) za{kÜ }THK̫u0qn~5vMR88P(IL;+BU:_$ϣm2&(tO< ^:X(LdMKa'W`,/[9vQl:0qf5>$*cqP^ $J6|kR&ᘩy=x >.Ow\??H!1$V͙9$7t $EkPוو$lp!&*̡}E;':R'/]z2LeC{bf}ʓ.=҅;uUP5ɏR?q41`>[(KUIʉ#iEќ>!+Wi2 t%.z-Mvt;.Mw"NC;5F Ppa2ˍQt~6<˪]axt}zї;1=Hs"JAvѷ$A4&=\ O)֋gRHWwc캫+,B8)0\:V!5Lb/5hX8 ̻um- AZR<~qíV$]Bah_|;Mxǚpar;mKw n=fB j@XJE\N&LjdɋDdΉ @~.G T  /aQת$}vpWRKh4,UW\Z.?Z0xPv jp!MZ4t#n"y|Ll31VS"d 3uYBaԐ7<4S+oEŢTą,ņeWC&K=#`4:97 .Sm/"PƳ~2" G>]g^۩Pj›a#sQfYc74 *~3S* YQjlІ%@B|Ah  ǷꮂpJ^BaՅsEܴ}B@IYjc)rJא)6SGY5/ƛ?ĩiXNHHSRW\(`} CyA?;V^(jb ==+U&W GԟIeX)8B`^`lGזhX8ti$V V1| 'ףN3*1R1L CLV;#FESC>i R(UP*ǚ_L>:nmX$/x`-l_(5?r@/dcA@ LhFXo!k|ſ?isTREJycE_}c`ЌTHuTMºWjXT |2R^+XUp@GL@NluDh'-)r͔ڰ X-lHU4"H)6*/R#`CQPثtFeKh4,_g|XA8hX84q%diRsKh4,Noc<4ue B}\f\Q( [LԿ#]T8%4^3Q shky1xH `T T0zbMY%D*J^UX+G{ SCk F0W9*@b럚.cq >cq%>C< YX*UvJ,TdпA="V x0L@GhQ lhV,me ZY;VEk۽)I [ΞSj}xHDz2\ZarZKmt. f Kt%4&yj&cA Lr2&yPd7dWh#hm(TXsl2Z$)S I1%2֎iLۨn}ս>靼KP\V9Vj"1墾ܥyzg,-r/7"\ݷB|GӸx= 0Vh8Qr |U#6[=<!F{Xa1Q);Ëq=;Ӽ|_-r{v+!3E$[J$75I{icǕ(x( Åz=G'00O3 W7~d{ü짟T,cI½A8H`;>~t\k'9(.o (|g|(7M {az Ӝe#=c`X؆o[n{+mL_Nr$  sl\U1Kp̯`qX;:$,7R&UaNDqޘ1s17zNogڦ mzE"hXJRb H2GtG FUuɨ\_9=alWSgp(t{r m{pǍ)ën؈Fһ#{>!Ƨ_~w+ʁ[cҀ=N)\1VƉ@0,>f,J/?anu%`qx8 O ]>j7/@y쑼e7)<u6c:5^Xǰ`Z"p?_fe);Sl1Ln;亗㉱:$?{鍆NIb&c1\DE7"Fq8/ϪHgZhZ JoVBK 3<ט`p5NL i7kr}] ROTQ -_ryRl<O hDIgkƦņfm*nbH !Fb ]AZ,'a,'`3]/1_Q.V{_͝%p< 3-B3'T7b0'WӛB}Z# GA: wcr.B^=sϩoz4M_ۦ6zx6LV:IO@ǫ`TNH~9IãQC;-߻I~w:`4ߕy"0=#oF|;3}xQ:ϐGo W,ۍٝ f#M{.Q(i;8O`gT2^vx^-x%`s8wwNv6 v]kuc}x^`l,~/wznQ>|:=jO~Oo[wNsWgWIe\PZlpzeTXI= ⥵ bLQ>nmk_jb90.5! vG!o L0acpAbmB2uaw{4$t&gҁBg fE!P qRyfAŹw=u(.4n4!]( U ̙`aj(3g "mgY<ĽhN߷cQ|a8FppjnJ|Y| LI9TA %Ͼ:>>V.i&77È 6E ؂RAWz3Bdi;K['׻uhSBF"0u^8,9VLQw`,wFY-])޶~8=>k!U=<춿o|)<ڻ':{?>Jb3QE@ #@*i*ˈh$Hb\"3Il n*gد^4-Q=Uh23 Rڀ .2+޴ۧ?:# Se `!)e?tx )@j (yk"WexCggq>QMw]ۇҖ>iVvkje␆pK$U, +JQ%W ߽G-?ԮElYٿ(àu!mЇ1 MtCG, [ُv7v_;…?>n城}?Ȇz=?N,-r}-DPg.y\kww]~ ާ.+jqt&~xHyc]/= pPz]K\<>њpm~=% {v=sًZDl!⩑ɰI'# 2 )aMS"͟p1\ `d2-=cIdC%C2 Kn^~sv| 4pT~r+vNle>s+`Ҟ?؝c'^؎,=.yDwhaɂwx32wfΜYxT 늁,4U~D[骦3I~"<"FJ.2mFQ19/#&mH선OHAS\r݄H6df(0*NCz󔝒)9I|&{6H|z蔞AeQ~(r_"[ul,#~q%C[:qɂK86֠1.z~{#V.557>Xӣ[L\f|}p6aNUMY(0l+*Gf8UM( .6a^eӨGUv0?VLKrX< 4 )b6gWNS}L'yŊ,w1ߴkf" Xia¨9q̠ MMY@bڲ{R8TVTR7a~X$0f7=*; OӗjBIHDUݵumW^}l:no!z*3Y(NQC2ZezF/<<.HO) %9/oC,SK܌!fq+T iBM [gd]X#p+5$Klkk M]iP~8hue6KhjªcÚ⪭FZ5n杫ZsUvBZŶdQ˷ ,ؔy@FI26 uզ[iʏ;:[뽃7h;!T7PwNvO0Dղ_8o, IYp"X[-s}+Α poZppa;g;Yao۔`|o0/k3Ξ/z\)OhnEjp Ga5 /}635ml~vv')f Ŀ4/ ; ꢓ.ێ@ i,zFV qTI`)VGmADـOsEiʣqn=G7= @Oyco' qĭ^b.+v t6sٜz?P@xM{Uh[`S3!C˭{ۗ%2ކ+@'4K\­:pVjl "zm߻Dsb87\WF1q!P'5J@;!1ynSP' vҥ,8'F7nŧ}~jA9f67f޺RQ+8ҒW}tg>F)Ȳojy|S5>r/kdn O Usnd95d9#_/\y×9sg֌Y31κ9|U3=31T-+9$}[ AI.l7aKҶːqMMBU'Z=&)7%,Ň:,Ƅ0yjI(R S 6GÌ+S^&v@g+B>#I[ۂ6Yݛ(qCr_phLPv 7l/*J{7kq%qJ&f̛tCĝigRAY aWYߋm*'fEuau̻$|π/WΘnp rOZ8,Y?z'M؞x(4>o[5LcgqRItڂ%\!wA,ժܘcIӝ'YVc?/(\xVXe*-sI9)Wr:um59S'Z)nZE"ȳ6?{Y[wQjhZ "Ȃ_GwEG'?K6Z}GWdìېɅPumZ%BF=`|`3i Nz%.䆡 as3޳,bj)駖AS*jgJ^ZZ!EsgK&;.TųD ~EE=KER 0.E*3Q\M*RZ⠕472v"S $Al:}}G9wrF&i>.|^* ņxK Ne욚8*e !ɂ9 Q42Q:eT.kG&o=ܰwE Lm~{4rІ\@.oeN ;B><\ʭ͘y))Y$URڎcI$) OR<*ٞcZ/-p? Ν-2x{q z?83E[| _E7 RqDSM) \55z#SQT`TtŖ$M=¥j|wLLP#`WtސqRˀ4~j* #DѦ8(puUj5<Edb{.D0!&6wa똚 TRuf3Q&dAۡFE,[f28lRXcY$4U,O7$!#9qML 欬mH+; 㴬9洚ViOV7d.nU= PpEFRA"=JȺnlUK&:lӑPIs-W2%؅kԶڪBU4u[uKEFuśUJMB'&BC. #rw:?E{~Rk@W.ǻ3+ /?ݽN !nmͤekdFk^]=[qm]{FٖlpTu|?7eۏXB>rT'ш6ʸI wuAT(5d8dͦk7r8QPe|s*޵L1―c:i[3Ӳސ4ϵ8U9trW^Xnrx.\xdOc=?Y~1(^!!Ijj>#ِq`n(=NR[['S ,Ul"X翔TVx=' ` B19%-Plj1ٺeEm0-ŐWgS#9;\ul8E Е+$=SamZ=06 ba,?ąmNC/_ aF>C>{ wbv x'gr;ǿ~89}Q,@Q"ݽ,~8.évI׸pzNOBD{1vs%h3f Ȏ:} v(įqޝ<A$G
7Dn c¤lqB#!NhSE uQءMCQo(׭ۥk귉mo &<ȍ3؅;h]tX??hӎȠ ߟ @ȏhMjK|-\?[  4܏a1ԩyƲ{>Y)"h@QIB>X``-  | 06ūX)XCBI Nog6/>@.ReTa?".-Sp31`D`(1| pr6ؤ$cFM^M$?ai+ND0l)>>z)hP$1ZV8$LeRJOu$4"օ;M0W.Ǟm6Y$=r6s*4N1k5eR%)KFSmJ5BpjvB٩W,(Zz{/Ks" ݀}7 S^{xxO<_dCsL\&ꪶ`,ʺDu$Q>HhxېH%aBxsX6KKj+' 8j[`QWVU?<4 xO߈W)ضb䊪CMK(ْ8Pe:EuѤ uU3ujidP&ێ=#2'pw I11v2~\ܓ-{UJǟ_cIӈC)ς!dЅKͧ<Gjadw7tcݿ;hU" pRns3uZ 9!4h 2=aF<ʄ_ |n5FzVcj}].}nj6ŕx?=]l܏XZ=JOn,IeE)-8\]q8CpVj462.p.dw@2[Mb~-lJ~) M2a4zO/񧷸!igi۶+|~M/mC=VcO#;u:]dKe'N& /ˤȉfZG"!`X,9ۊ*:E#[-xc pN,&ku0u{/@twM8Reʾxy ZuH6*J>2Lx 7a:ByTp2)0Eq$\+&}O8!۴k'Ypݎ7>nnڣE2޻yXA_w۽ljyj|5N+,v5ݞc ȴAP9V?nt:~7X{ӕ?#w/=<zVzWvVcvnZ_aK!{ۑɃ5K;fI ^5rsEءj{ w@Rl'wrr\i]@t F`>>?gjkYgF*hn"grBKn,sUC4J'Sr?XšLgxL s.cW= C Tܥ17SF9SmY zzv*3Z- =Ζ61q@l-0U.4z5玱G§ٙm{LKM>\륙MC$7l * lm"E5LתUq.+ݍeǪk+r]ܖu[76=dB7s_3Yj{/["lvީ67]\D:x]ì^s>>8ڇNbÿߋѴvZ6 Mm53ik#4W/B'ϔ3np돦qj">?UCݵv),ucE=cr… sG$Z$H$I*4 dv1MSf>` R1šd"&ȗV P%0l ЗZG_ BU9ߋCgV0* I ]keս+[mo%ɲfnmoߤ[d7w'W68 |$4P%81i('!"qcFU"=P,ʧ0'N|6 JCf%܄&w/\SLTi{e I_ft=fGE~Dؿ&\5I4Aڝ+! N%O{=_rg9e"cRp"\߹+aφVJ_V:;M4A? mt5Q<0 .Sl47=Rx未7o;lMmW^eʵ)_`%9@Mٞna -e2|X's.;XBE|^DFC|.a3-Ku8-6v˓n.Ȁw41nDwقvuEޫ 'Z/p43Z4]ٓ8eY'5$n cAF<kS1O :\>G6G{eo *UӎU/"LuXb537jTwZT=]B0PKݩ7M\tR~Ry*)ʽKK/ ϓ~_*u/p.ْ]ng/~rb ΋L")} >?;\afS,a2C :rgU,xx>j6y7tws ZOԒbJגZ(M4d\kZDBz(=_l/e"^r*{a#/'>uﺟɨprs*ṱGcDBl30J\e{3y kVO*qdK.1ث8vG lzDڠCRۆKngvL쬓7pŅ~׬b~}~Y997'VD~Dc"Ah( u9fɔapG V12_)Lt$!C”}ͱ4X8RՁ*%S:xyE !<+TS>PJ(3M81ЛRIL'RıQ/$  Nб=Q1'z+na9|Et_}v_ä׽ﴛ&d cv~7}C.iW4",`XD5lkOxtqqP˳Ɵ9o,T۹4h~> 1=kEtQoq 5..Mc~tvQ|BrxR%kZ}8/}{Un]æf` ;l;-N0%Y*ʎB~I@̹xUe- W VvݱaDNږUR8fql͒00g(ov9zZ 1}}ЀNn9emփ.꺭f,jdR -WJe@Ig1ْ3`Khmr` F bNZ05&F3g L15_X^:{-P1oNz>6}DZDGV1 }m5[J%-.'͛ۏgXV֦q{1bN7  *ZV[TWm'--͛!"4dw:]⇟} V}~^ ]*t5BwXU-?Tхxe4:͎UҦh7V]nػ%x$TK8\*P"&.7%i6)RFi.\|XP!_@@@'xs8HʐG4$ YE8Qa ;.p*ڣ*HR1=̰(A$ƘҀx0Q?fI!τir>ޤ?['wD&vl7 ƽIfq꺬bOt݆ h>Ei~D˚ ;۫M~h.FqA֣Y~/ }ͩfڳHcH_$*`DjYB rKqnw'W@{=Mڞ>#CaQC𦪸 .B3\LOh#nw3vYؿ{;0vSxr$ҋ[T xm}*-> >m6dHo 5(c~KPGW7 XNh5#m{`z+U'>RccF&Vm3-'=Cr-/ -2\[dp0s=޳u^ߜ[V P,T8\~N+v pN̙ȼfLFVsKru{Hfs8B^.?07?p42vcG(n(817V\OL&Byqt@MxD cy xG0\sJP"dp:[Vl5>[V5>p,ܘqyڙc!X"'r~R߿GgE]wܸqFǯ_GQrX3c_T `1Q (B:LBL{L#`龞iK(%N4K31\jG1 }"(Q{ q!4  ael @Ntpvz8;97:^bb^ eb4!8%̧Q. ă}<8:?ytXyD1L`(HVr$:I"Tµ"P}O`̨p3A0 rOS m /G}$L x  }-F23}{To={syan1B~@ifqD'H)ǒS+R <`*I,go X_q|X_Ț1g1uǔa%=2I(¥dDhffù X(AYュ04O3A#_؆- k-|-4]ʄ$$ I_U@x]WGW21܇6'=3= $eFƄ9tbiS}VuUׯӨU!Q>TkIys7\QZrU-0,*Ј tP `.|V==mpުr%kpKz`7 '>T=qԵd]ݗ XsI\$P{,'@)qE4 L Ak-+2esa-^rl /HYh̆l/yqlY5 T;Ce'&En=Ls!Jd)PW9Qmxc|V.ǢPOT*`B9b3 l~:SjOMS1T9cbzB^q#"u+ cH.ϱ+xpM䗤^}X}@'Z9 (ikR瀞j} O>].v *mwQv8_o2*m_\\,}VʥD7ôb1#{|G3i@rQOl WShs:g/!-݇뫈 D[3݌I+qs%.nN{6rYE " Sc,X2``uNd={(2t9rP&ocy O?2cJM1IeFtm!1Kq`8BƟq{@$i@Iꀒ]guW3#3~G{_ER!%mQ`M,$<>Z4dDk4^G6]}+N1(&GNHIn؏puc3H` W|z@BEwHɵN:@wO>y3FKZ6,^? =nsÏbބ@j= :wIOO:.Zbq`rFa|҅swt ՂY|~BZʏv_s"H@n.QƮS<:;P98+}2Aפ?^.dGĆ&dfG~ a:?@۠#ѱ^7;h50/ǦnWW4aw8ڌٻꎳL|4P$q_owbۻ,"wxBL]x-|ijv0>(=7MpAP, Z:;?`DffAbG;Brr*U42  1*ˎ9RIC Z0W7}yyQAc )[P4[1™poLJY <٪[#KSJ᷿1 ]c+ٲiK:j*LEV-_:bɔŘESfr{FQv OQlTE=׵MR[Uy (TN.0(F.uL`g*i:a=jSAvbYjxFm<AS*i.A疔X8;qhzնc1Ǭ>:ޫ> 2>2lqXCw1 A J \~;7r',~ew^?y@MM>T<6V&Y䎳5^y{]+>qj+X|_;MZeXro1]=aa5OG6ՋNˇ2TlHV&߇cd6R.\s!$I/gݽſ1->ƂHƅ>R=5Ψ)ŵLlW YE8Y?j[ څ[o%Od׌̴L||xQJrfdMyd9ɧ/̽t. ,}gYq.)C&ғ'c VG&^$>}&d9˚SgNիgğ( f,SmN'"Hwe] PQ{cݏbIvs}|iO?'CmeޥΈ lx$R,49\Xhtt"~0\HJ&-2UjeDJLd= aBXFd$J7l+' _-W2*mHfj8c^FmdKHڽak'{8[{tJIOxdt.27bE.qHχ $OrFx!>M _Nm_֠'U>T H%%Q&'Q?.znT.cZ ڋύv9BB~/x]2Zf^/y\\&XtBi2l<dMFo~e9rrPqETH>qYV6Mƈ`ߏb0{/䐸E.n\Vp7C-)}utcm." K[:W(ʥNx^;q[D=VeT@]Q=`#kj#\)9`燖#p/ӈ;S[ /P~qZ|qT[iS%+<$YPRA1>4gw-ot.ҡ-f!bkytѹƓZw2P&M$﫼I}UB$Ҏ,oNRqo{EcTu|VJҰ$pgR*䧎lOnn3voN…MW^URjhW̨Xu]sy ôaM免SVzw(sg^JL2I'29zꜷ`k፼߫"q&$Pua&Q^=mF況l`>8쁉ped .Y:(&4ZW'D ᩎTّU6,/aR4de홎(: ,=pl˖]*6- (#Yrx7YfeBqMJ?Bl{{cdy։e8q"WqJb][ @.`(VkI%JN}61&sm'ξX5(&>>^)xըN1>s.ho$/o^DwԎ63:#o~")zEDWN`5佹5Ey_݊v~5pۘ ;B2; ?K)DqW|Rxx?8&m"7xlܥ ȃt+O8DJ^O.we}yuD4=K3pyšu#&xR<],+bk2J9=#'v<%1GQd[9._m+*!25e!i1E3tj>~0c):P0_>螥km&d9l2GOYeYi3tH_|t9ڳiE }xaQtƫA)>u1}e@NFɚem]`}Ás& ZUnu)vZ+gu,Zhy'sjݭYiwPhBt=YU*s{X 9~{e p+Qw /{?`_P .=oi^{ 0p\t 9Kv% @."R=vFON{~2 }:}py"~;+W>x$β*YCǗҼy=d 9J[D @xxmy sg 'Sw5k w("fI3NwU3|ʇVm510F@fE5/s\`7ۼ++*fE/p6Ob[Nȯ\$;VJ]E'0tQP><9\ѿfeF[YTm{rY*j,S]M{Wy,QPOcݖBfj:/S>pS:o狺:>_ &kB~7 ,f#ǵ2ME( J~|erpL<].od'%z{5.v곹0<Uݲ{㜭\Y2Iȡ|Y-VcGaMKER?hM-_WLۊ^YIq8+'z UUm.·\}Z6}dޕ$gI.lbvwuΝo׿` o+{K,[>^ ˦tӢ#@ylDV˔'ڌwe'EK8*s L"*} EfOQ8Blo'bs/?y)qLP8lJn;c6%zLVRG ;0;TR}k=ծ4Owu(g=]nz7 f#[*5]~ʚիQVMzzBpO:q-P+V DHڐ-_"syl%b HP"u -gԋ `?e$)azPՂnNNziM)Rjlxj.To3EEsɌ_EЎDt` "5E'H_2|LC"dhovCһ Ul;f#׀[~C@IO:I^&HN mjF1\cvThQG=[ 4\ˉ;^cJfploțEC;Mk4H=tak*0kևK]o4>oc)kcrZ<_Ae[\pam )ڭ>5+H=M%zD]qTtYnusf,d] dyw꩹nK%#WO-]Zb5R`C\+5nO"`.g]_5+_¼dY+a]/c_dyqcj鶻WێJuj-k ZY! eA]^d8Y *oOc.x9Mg}$}5yz >E&xѯi:{_aU}vLݭ!߼|[oInA;)hB[3\Hk L8Ag$o}H @@V`,!ÀKv} >a"R֏"Pa:6##,o?NJҰ@U$Sh{iڥ{<6|q)wJ |xw0'Ռ?G+ߖL@P(k,br+F"D 9'* Be͗KW}Yuv zr9μH'Gs gG Is槝[i| fSw7{6#uT;U)ewJ.DrnaXV;3PHkxɬ|'C ]y ivt_d[޺JnWg~pp!c=tHUJ;R)Qa1ӀaqB,ٞw2D )"uh,&cM\SkaR82m][ޜ~k1QPFڈHĔ0cQK-E,\kP2Db5G֯~s â07QxW,Gªb* %+l,߆|Kk CKaH|?KC,>,a.0jHس[VXR~99:gg V#0V8"䡈b eISt]I\\ѭ F#LG;&%LācCP_HFƄxP_p\}ۛ bCZNpzdmꁐ J 38iB/E:,"X|y6~T*tew}@|_Q02^vւJNgDt[PZakuۯ4[%7:ݕ_Yl~p60SjV0Ns6W C5ozTloL ӳ+Ȳ㬋D!^'$|e%(S{/t~]F{RzOJkOJ!hNC?Ssӱ:Ε#3" h61߆{mT;3'1|NbKXF GWx^ȆW"~i,1%ugyp6&g: `[js,W|یHRm a6 2.pp, ai$1ȴBLXCK]Hyb9riwSlVhMiu=xG^a:`c:2cp6}MiG6#zӇ<*'hF{ _L6.zMV,R|C~2ݛ7>dU~K8CWνcax\놭۪֯.]hee*/'(?w?3sN@yͥ+TN`yw#]!vcUG(CfI˻*CF)Cs28r%kӗnk>T{ˑ#=ݷ<}DkN˂ϗ@gk#Re4{<UF3*Y͡2`ܚOw\j/IA=(D$ QFmǿ88qLHfB27oȕ4ZʘY}eL2+2ٝUqwRWinjT rRp&,Vδ*v4f^7yOr k2N"Iw'qټߕvEnښ^OSş_Cħ(lɲXީǿĚ%s;4vrH c\lQ6J%H"~ۋ v`ouՂ\[ClCqLq_|cm,S0]/9!忟N/A(W] d9hMm#,kP s1v,b(EH.9msEΡd (lx*` 5}='-t:K& USB EsTeJx6 4+=ѥvX.hgBE $yV̀j@o]Q?B2.@Ơ) yjb{+$R Lh[/N_7UUN%QKUDE41\;ejy31chdr[X5;?*5 xPPZskߖiI|OI\ HCBP:$! uHCBP:$! uHCBP:$! uHCBP:$! uHCBP:$! uHCBP:$9TrDBl;N3Wc4kf3$y$OI)yX<ĭs@gÉTQXTqeA&H'&l~vSQrO=Qpk{߯O X3Giu ef; ϫAu^&[NldVogϾ.GteG\Ais)eً̼kK=6N?7FE~u}yySg>~h=}QQڙg{ ۯZ\u\@z~cPfq'ж<h.+Ѳ4[Ss|lM#Q"='!rVy{}|}`4\OfVa%CpL۶|"fuF+9?vu,]]ݬ]ټ[iܜ;^,Ozbl2M"7Sl>qnxmI:K{2X桅+vߦôExwX~:<_iP\fsM9o[̚"t">q`QPrP(&4Ve҇zTI 60SZL// 1.ӧrCY8e]n'1l_2 I-%-6&-Z{6ce/*jUĩ*!,ɧg7v D.N!Eo8e,ÙQX#-ɼgGJZ>xfdhhvdO<)bnX7b55;NPiYg$JD'Cuj\M/;.7! EQa|[pR6 #7gnwy 6FА޽!wC~gyh!Bns8jn'Mt~4 OaeI o߄i sڬY>u@`^k~|-aol_xO~[6~'UqSǷ/l2Ƒ%"nBZ) l =sIT Rҗ[x3<djwE ҭuy'ِBVFZa;#6X_E"yWLV0Y",8{LVF}!oF+qdB2ƎbfvA^W2s`#l7[rf.]z zza..ik GQk2vdѾQ/ ק\b<ѱjOm":ֵ{lyoA<\Ujy0N[}Vǵ<ؖvK-M{ŶDsEo߯'S~_f Z͟8')Wkļ fk4$m ?q*12(As(H dkM&ߔ={9*9e۩$,@H)i+HVIc%4S]^d4~q_>ĭmaP.V/>LreΫ.淳vOΙƽoxnq2|$6\3kڊFef>bW:D<#=.i 27dl~KAg]P| ZhOqpUBZ8Γ ,*򆪰E>#-ߒĩSQ Vs$!kQr%|iX0+8DLyfNXO)1ZYA{采}7e ?G7>,yrٶs~HGbZ)!GbMDEjgA+QAܵfs&i'мT׸MZ/а-[\][vlqXȽSM8?6k[m|}|冶"m7à-PPhJ1RЄHv\w5iUp#F\p F;!)AyC @AlGڒ ovU??ȕ!>_Pv?`3Y,s]n3"Pl#5E ~$Ƀ`Zs"JaKA76$u OEMJaRBrVuHPPsId 6*I$n(|~*~hЮ;̟^vZ|5v<5:7 ;O?GQ){6i8`#U(fQF]`RyɝAQTk(7|L+]䞃Ұ$F3%fϢe {I#^%?XC%4V5gбAǞM0Cg$ Ӕ52JXC8KbVsm ;P4mдAΩiְ3(E4lña irFؐXt}9XXr>X|Ԝsu$OV.VoފtCCc;T89X/_a7#Xڡ:L{NĈv!ah5D,=&lulR2#^ҁyKA9/Y?RM09ƥ%*E1P)YH:*<0͌Nalr1μ xCC#{-IK{ECAdo! |q%)l0'2 #ƘdHdlH(iSs5$Oj MAZP 5tr@솩3z6r@HIva*Eov2egeяXR= =Ý!% `]@s b2?vsXʭ\=s`L^_Zr6Ւez9aeF&ג,)3|H!쑹ý1W]b2?XኍcHG֓K 8O:pDi1]+djqm G \(qg52r8j1n2ϧ٫e)^F\ vk1Ų.و$Vn02D0ըU_Ta8;({6f̔oaA#%M\5:ݠPޗ Cژ $|1FίYXZƁ[5rF Uf~ˬil!1[,|Ž^ W .TG T~yS]Lqʙ.,#E;ZƠ0?Gy;NQl^USaqŠ^,c܃l/oCYј|n Ȗü9.]m<>f1-\媲h\bgKvWpNqw.AhA(ךZ Vϥ9Kιh1-uy ;ӓ(g*[n|ؾгy"ν($eq6,bxcy6y{궾~$҄n>]z3 %qcjw=v|9zte`SmOŗ_Yw?m2yj!,U;Փ4{J<ۦsuWK pWl niK˷S;0DmY@aEWF]"ZKj$p)'LH*JĖX]7c)]^RSfLv]@R %{BKߙԢZPrҐ.{B[%RRRݣMt&}R72[M D5D\e鍹t a}1W[.Xa9+FvO^$uIVNCx^$>䷢݋ X=8N1{?X)hOT)zB>L6L.,͚˝?eF4|^UYA]qS tД%G;j:P]*f1 &-#C V{|u8C< &(&r$d(3J/%ʼn PpF98:ۈoPxMI B8#E-c}%7+-ߓi(\`emķC(@<ǒ&0--F1i!ƒ#mf :8N|%'\{⒊)A-C 5%DiZ\'>HBe5c$Rhh.J+(v8pHl5#YkrZwPx.UiWG f\]FspڔM[[Ux ) JX+!x˨tRb/"DD㈃3E3<Bj.J0+3]"jG $ XrljBē,S rk0WN!>gh  ^ Ko!'!J!"9^ZA]qB|%7X;(b |c@3^ќz 9J   ^iJ62Bt$Ǥ-\vu;H&A\!hVҠZt lJ[)IAA<αPȀF,ɲWB )ä8 kHH| 6D'k#B5ʼn R eDŽSX`QڈoPxUiWa*1 a VQj,|%Z:r!.3- r\M9G4n@(A Vc0P#*o] ^V e*c@WCJsZyN`Di ! 1 Qb@(Aƺ)B4$$pYdwPxaJ2`9 d\Ŝi81RwPx#tik/!0q{` FO /1)I0V8Id&ArQ{-~ ,Gx ~y*be:4^/m!Zͥ7:a`RhHTBJܲ ,-{|oBnM8{)ZxHMW^°(m!`2y.Ā$xmTluwPxN+Nf8%ȁ n]i mķC(BdMze Q#cEJ-QӖx ă. @pU4 i 20,aV@%Wj80cG6AOȩJ)D>4hg3 : 8] …I PR$R.lh. ʼn\g\88Uxgy?XC"됃%RRĘ$hJ廧{A@+$fv;&Lf% %ĭJQ9JVBL)|:)]{3ƅr' \H.˦|ۇ ]^rtΩYrMe&D1ya!Oډ-EÀ 9`^ix3m˩R{l<y`4U"LYc :ɽ'tRxf^h&}*08@>My^8CYr<[˻N䥐YB`5(yQL)Aڞź!ON|V̦e`|_jZ\=Q](gOLV1'h-8 D,ܗB<+u`2eCXC3U楁DieJ^|:)<_p|W"k:5=m> 1延|5-K|lK ^BC=xBw=ș:gzR4) ]q>ՅݷX7 ]g 0='s5tAB TX9f$!\`.߅r+Tq*Eckxp#!\`sdpj:P3W#ĕ~@' B Ϙ+T+ l9q5\9hL(9:FW Tjq5F\yjGgv++T{LB+:DE}ǽ?wJg\wZSJ5ѕh+qu9kBV{2BPuBp*Eq%g]t,Pdpr+PZwNeJrg9`[GW([*N U*q5F\ !GSZUި^7J8]b/C?!QN(!ܾ[?Qaqy5l^wƽO^gjtjwg>auoԷ\}}q-&0݂]N3|oupB+q{M \e\E r//0ҙSGn@PXLWVJ=!\W,k\Z+T$q 45A-h%4\S_=WWW9cq:2K@|BJG\\@ ZXޘ}Vxp<תongPxf%vׇb].R-S#Ou_wUuwGMH.Ƙa_dS5^`IuX啶"?_'ֻb3O)ԟ+x5IެW:엻IM v-=>xl5˯MmUMMi,P-?I_ a5jU9H*^,|W5wW]6#ldvoeWM/SNg<`>uө"/mלz ?@?XntɤN M>Pd>4hV+lKk^uclmDX7i=}هb v|GYI(q&+-DM|i5/GdeVif wSTFw .7ӆByxjf°%MugeZd;n jmbSuzBOX~>߫l#2[f믮[vځ_%QйgvXWW_eCvKҖۡF W%7QAhX^.,;A21&1.X 8GtwY]V&p0o y_|-5 N7"5 kfW J6d:Xcw|C#g)~xa!Sf 4'w]P~^U^e!/yANA%9  m:9 T+Pqhs@B !\`ҙNW#Hs'BaLܥ \6spyA1W1cϹ^ ɹ.{Z}2ݻޠ|1FGƹr" t0(גIZ׍FTw2J`!\\- պ'`@d:j20` P#+Pk:PW#ĕ1VEg[\Z)q*"ƃ+kהY+ P4C,jrpM){Y5 @d1w5F\ynjjV]\E&D>x\JsW/W*jœbHuBVjv*vj+qu鹳 g v+S PY(\Z+Pɹ!L2BK2BQrBE\WRpXuy7qw_` 78fE6,N9!V?A&4GQ?QC?AacVp~BE\WiRN`-\ܳ^l+T)Dqe2gJ=W :B PScĕc;J >Bpr\-SGWĕ*IW(I\8\Z>`Uʘz9ҁVߙ;Z vo\+{]Skz:"ʡm-p#5=VkMW EW(Wq*BF WopBlِՖ P T: qѮ\aual:2V2a̓Jͣj/)QѠ\K&A^ =°ь2Q[! ;IW 1IWVWR눫JKɭ!+,5 PGWRrq5B\i+l茮P#3j{[N#bS v @b PC-gۻ?f7"ywh˼*)OUgBUYdr,{w|O_ɮ_*cr=,u|/'g}uU_/{U/zwƸjJE2pI [Z$P%:6țK~Z\0|폯?~˙(/7CsQ[&ɛU]'?r^崧 rÆ}Ur<j?-u%e WI9 D.b_k֋}ˏUrsX&)8j8nZ4wMpnC0W;#Y]yPGwCvKs⽆PP-|)ǡi9at=7\Mcqg~]ч##TY0$#G眻G,LV1Y|v\ʺvjMO*S϶Nژ<+쉂1"0 Un=x+۹o䅼bgwVp仵 H4+ ִf"gحVeȅ`6 yѳaw7LVZp+\ȗ>VzBy/ F8CjΝ,":/nTϥv}d>-"Bnӂ6]G颚ok[৞m8,,r@OH Lvl6k*5UA2o+4P?0ݡP+%'dB =DC5}6̖ꤐW~s9TUϮΫ1] 7l yDGlTv6a?.t.h~e˗.3'Q ģ'k85h:<ݠ)I  40!%v{O{Z%zZNXʵ:\g!\a`\ kBC"ƃ+!Pdpr+P{򫯎T`UmEYp嵧5BNʥnƸ:PЎ:W>~ŒbXJ0}!C;}ک=mhlt[G\jzn@0g PnߛX۩jBC;?,VFit%+\Z)?BJF\W͂s2*M$>pWI]6!eҿ!7ŷ+ג4ܗK{_#hz8V#TʎSVVɶj-)rv83pHJd=Ճ>fn5Ȃ wl%|(%r'%*'TcH:M`7jy4n.{4<^ުZ/:|32ye#yΠA5}rJ`WᲇՀ Sv>y_ 0l]@l>"E7p: nQvk=$}jQyajTG󬴃6DݨH3L E FMmfκlTхU%o\[ G}=K:W.xb%=JbPsYz>&6M)Q9e:O)O,#,S(vJMR|"S90?*;p']?+i,JVtr6|^)l`]T+\]VkWvY )"UDg ; B1"r奪 FQ[RJTxřX_oTqh8?F{maXiX͋N3J;Ν(^{_\ iן1)Wѥ]>J/&>tHļr*$33i/ )9&% ͛DK٠Qil2)|YeOVҫouau*vТl3m@.+ݶR"n֡6ER+0ٓ<{d!*_BX<GhADh{鼄V&N70yQLcQ)9ɼx<4N8*4ZZ:˙3Μ4ac"{wmn9ѓ3- CiL UNǒX +J9ûsc”!R9x4\JOQO)e'i$c9yq V,[oy;{%kl[rRcT ȧ2=^ʻL_eO}“,n|-g/Ƭ"% (Lؘzece lSjN.k,[`AZ?AHܯxʡl2 _YV^J0>8Qe'_1yX^+'݅uog<қBnfi۾h;0*ЍWac; v)>o1*L˖ z3ͣĩ'rlPK$-3P%S0hF\#F5{E2D?(Ⱥo7a<ݒ ,4Rf2tSq.;'_+=.eܠbpYڡWXRܳR,CNܐՉCvT[M}M T =Ͷ}ظuM@AOW V1Y`6mН8L ):/܅N<) &>[(VݭC,jýi#[7&dX؎f[T-ދ4A-<MqtLue~RkDN\3g C剡|Xq88!6SG /lJܰf3=^>5YvlN!Hdp!HRNcU4Q&E!i|HvR֫R/^sʹy,9g5ED[R0@B+GLZI$X>q[fbv=Xj=]C.l^zfy5]g3_^ OsY)g-?pp7:4 l;~8}p7=[}tk%۶-߳y CZ)"B& H <юS -Ma5xZ=+-i,aa{/G(yQF Gs׺-]HĄs$ +G6Ggh|;yekgUhs&gsIK8xʟ.?˱jc!P-lT T0VGk3g3AR!ϹRΐge@NzWz 0Hr&3OSy(U9@*=b+LPʈotG)q_ >ڡ{N3B.>VG!Af(W a{ی#&>:]ߨϊ 4Rj mrb*PKn8"֐ dԪeˋ?^a JpFi 'S4&*-fIDJ%É)vRx&bjzEx;!BdB렽  @fP 6LVHj\TnCe "X2MbdcL -5 Q?kwcgnN:ZipKz,w &!le}v+Uf|Y“!y{;FC@ ,uOMĉL5JA^4yAo_)lqTHsE\QailpHjKB)ζդwD*kTT  ΄@^jl{)JBQA{Q/ek-vpܱ;*sD~bw3xr5=DQ"@a@h vܤTr,-L!=Qz|bf/HXN6ԥ'8)'#jBm:4mдAiOZ=$OlsaqJ߰>>F}ńX!c WV_ۿ.!t˾5UZ+LO}|.kn0c?[9wDU}J<>J5B;LiJXcfN$D2yBYLj4J^7Z'.Ώ\ѯ f9xQwb17gg&3. QQ ?xV~>}?<# {p%eFF8R‚p(>6!UhwUuG H av4J=P?BB ''!t`r _gfkjG$&poRŤ  ȄSeAi#HEY'1\93 č1t?;:@.HID,|4Us kEj=s͆_f)ĿFYU/|߾.|}q׿:}S<}:΄Sc;HH3 ?߉=7?5gk"EuqL_o]3|w>_ߛS٢ҡżZ*!_dQn:*rƬuf#5ֆFyԜSRuL`.liB]qAc)clJuT<>I` <>2Mv-t,sQbkl"aK#T~d2P_SG)`[:A KJSJHŒ4޲Tri'hhp;:TO1&=X ,*GR"sC*uzY]1Q|;FzHgXt'{u2v$h@?w) v^oVxb͵y01J2NL3W}FR.}BtLLc!au+Ԫ -~䯆k2C[3/}<=[3ۓ6!&w-qIWhs)d{ًn;V (į_Ɠ t a2ʊp*ϼ7vC;}\cWp/;рqGdxWz;h$]u*VCϧw.Kf kD5jGkB{׻gw/7K.lm3{nm<|W}S- }w>BX_dw. CLRxׇ- 쎆]w9X`e0 ϰ`ҭei\ͱ;snK&ȒIW~> jwVVn'>[ jk$}fۼF% "1fw䁃-dsukLii5{]Q7^sn'&rMOWf2cjm5I87٘_O?_Yf2A6]4!g+] 7p9 Q{a#قyvU7n=DyQ`OӠo\T!WmZ.k{nQz~Cnorެ^.CC³`bo,椇U~fC>ayAQx ΦP;qܳ]>;:>VpM8SMab;ʦJ.ng_B5͸u]׸Ƒl5zf\}ʽ\#"i F LŦc'AV.4~~v)":; f{?x/(($-Nv rQC?qf;^7g?}.ܣ'/9Aw~;9i-*(ei#]_E1D-(ޘ($v33J8|r`S~ya}|ՄaY5ћ'T7z>6zW?Nӛ >;Ӎn=N{)Ϭ]u\I]¥ȶ-Qp2{$@%ٰ'S~V$ ؽi\76ǎ%\me{4Ĵ~8w6. 6l'45w˭s'ZJu&Z>ꢉ%{YUdu8dX?ͣsC}JZw̵+ji ڊAkPs; [JPn)wcFJ(-rϭ?c24E5Cc!al|nVRjzLM&a~+ `}Z}$$G[10KKܜ=H*2!EZcC q՞:K+vo'0L'L˂BHJ~ ld k%)-ϰAEEg;@$h-]}YqDvU BRCV9P"H1:±e60>Zb"%l00:i&^IFꡃ} ģ R'4uk66J16blPѮj t{,|#Z}A)ʚ` (WJʝ-@F.E0 8T؄>wu[d"^I6 vk}~]E{jPB]ɳ܁6BoTz2Pư)0mr5ҝ`1g#a a11uKQ0:Ii688)ټ߂dD V r6Й|6@@0JoޛZ@QŁ.0є#/N0 VY22oRX6[gZANwWH]=ǽ( E}k FAyǒTq !1T(E(X(! 擎.$@H4}a\"µ8bC((,d3cNA=/.^bA]fҼb6(9yh3(RbvN_ y7C\i//*>Y/GBیZI8!{AyqJ_` ۼ*OTFVt 2:0AOe,0_*XX肺`z+R|&Q*L&zZ#t^W1`S,zGNzOyLP$ m>B:BpoNVUS]ߙ^ yǦʶMZ WBNC/u !tz7?̻>Y,TGsW0%V XF8;K+$^>bUoFe/T paL)ڽ,9jq[ GKlԎ e>[^F7HXJ+PK`ixU q%2 1<Ng$2/@GPf^c ơtV% ɚjjʐ pyPZGxw؛GD]{y a},(-꬐-flo'jj*υ h#M{ytՑpiF΂jRksy뭙+ .BX ioU! H> 棷KP03o qgkhΡ,£<_ᦷoHiώޟ+ әd@̓ Y0ҍ;yhl|IP6ܦ `]}"w60,:;5W7昬5Üg5@a[73Fʌ4Ky  0 9_DmN KAhH}bWTyi5‘rw(c$T:1|4#Ґ]ORuM` AqMicm.wLc@sz+fn1HZAnMJF;- &Sp pAC,!?}bG6x>58p1y|]~FwW<*Z F()`gAA8xk>x,nyU%fnaCؑH6]]@+aN5MI#r\7r |l=.=aPPNu}%0W L|A׉Ru^cP hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4sAt[ xOm'}l [BP[ ƠpL:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A AЦvUY>1hP5u9JѠu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A lPGͦv` ȼL@ud ƠNbqA hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu4A hPG:Ѡu{J/?'A1~7}x?x_H F}KMAA?~U0zh^3א+uq= ]M.l&Z>&JqJWYZ` re-pfjst5Q&UW5)u췳VjM (WIWi>I+q;kͨ[ku?]Gz/<eihzJxâ+JW_zꭳ! /? K[DDҕsm]0Մn+t5F:tJ6t]mo+Bs^!7mO E6jVV\ZIWeo.Eə3pN.ryY9ގ<8%W>^+ 7ʁQoތ3O;7g]q:{΃@PY/^A:Ya#T1R2 qtGdjiqy%e>Ή3bBe•O}&=:fRj3Z m|D)fVdY7m=4|y ,ŭAKjI^gʠ00Jt|XmRIlbg-j Ɂ#-9CN)bx)gGd⣱h\E=-&E -FU"XJ/jxz*Qɺר8ՊeG`m+βcQWڧr;CWQ] F8&t3JJq, ٟ'*iwՕd2;*tcF]%r9;u^]z=JL>=U"X1*|4CZۻzJ10M~ƽۚ9k?[G->.ǰFL{)cM`d͜qHpQQ h#ZHՏMj4wfjt#Ee&}Jd:i2p% {95aXaB [РYae&Eߴ!`aqވ \ UuT4J4?1pgM1Ƨj/Wah,]0a~-n]Rm~9lΏaʍ.26#]7rɯOkQ'Sʤ໿]9 ѝS_E>Η?-K]4usfF*x[YQ )*3r k6 +U'nŦB;7W iB^BM{܅oOn*K7Tzc(itU(;W0W3=v [ZHx@N`ZQbrO|Y`oJb &1G87M !( r0烊W`*f.'E*:3EGNR/VwS!2!C)f[.];ԒN4$] t7[ ԆR?DD;؏qf?r420AlF/aN f5EdYc "ej<qOjl'`5C|@|A_m]T <GEdJoP}NzDt7q_}t|?ӚikƢ핺9 ̒.䦓g)7~|An Ⱥwm,ȯ a|7'Km-(t"x砰ǐwA2;sͲ$ANX:xdXl u2qWwjcIK. v7P0limC% ݮ^b.´rI˭mwn$8FRRIM1 C_LR]]ˆ8kdڡX^sz:cU5.x`\G#ߟln_TQ o]crX)*6QEYEZfKCnμ.TZZm}}-2.n_x7 pI {@.+w  u8xTUjUe;u?}.b7ϽO9mg/hc;,tW1+g/FG5o,w.])!k>~ 8 EnbpdaIθ+;i|ov*y}Տst!DeʂR [X1c^ˈ)6h|њf57tr3Xyh&)f~t$4+pdZ{;L^wT=[z4u[tv[Iuhl&!/Tqd۾>۴MQU-vGfα]vުyQcYK%d4R-y-~E{X>M)-"ĠcPKبjѕ8[4 N2~Db`Ѥrǒ.UN|t1@żt1Zhxq0-hafsj▱_6Ty:PYDeYn5lXLjAu˰xmVt V IUp4R 3jg)~sWLgXJ;V[#ǽ9M!izC[_' 4ޣnmfc HLI!GbN5|/u 9ejS:;HK/PZ3}May8բO?u4DRFeqc*h%{gмu9ɂ4am?`7gK z;wBŶ{`h;V6VaPXЫdOPJN2ڏ?4@HvbV5F\@$d s$GG)V i=*' X AF4DP+93:Cz DhE^QCRop imwZظsX|eE;[ UXaؒLGj2;fP^ U 對NSj3 1 R}ke}I4R',kⳐEr=ny1 Qֲu^ Ԡ3^uԪ'.?LJ𙹰Gf#o`$b <* s]`XZʩR`PZrmq2֙OewvT~[U)A %,uJH11.yJ%bdF/cLoE_:I ͣM/ʹh=76~U`t=?+TȀ„T'P fh/#LHLB̜Tm5۳*$,r!cS5;$$ X M+>vs:dɘV8˄dPg($^2 YobA% 9Ij$f.hCG;I$CJQK Gzd%wG@f~# tw_~qm J2ČHI#|%og 3n <&Jݭ0.ݵ'6xd ռpf<%bLkl e &dyM>jb wd*2cDެI9މ]Cc,WfOsnk1DicߍZ&&ݨ\v|1b.F>ghZqt#8!Iǘ  Q0bOCY b/ܓȷك—1Kdǣy#sj0lܮ[ܤF9IxbUIM(9oRxhY甑5큶d&\.%6!`B2OJ;f2Rʃz`4X/5^gs'f,%m:/fN1a{m,DI!XZ)US~q77%mz`tv0-+5KX|UζV\ŔdR̢a lXsa4!&f$Ɨx 7,5: KwD@4:&(f׌; c  B' L,< RGQL+(X:Wlld[uƾ*"W|9KU_b/Uk:t++Wu߿>秏?~~Ϙ?U0RZ&N\47k*i%e|w6mlni|]-O_.w¥ R7gtMKG؋9+@6:Y ֩Z,f }B}yo1ҒlclVx\l JQ=][oXr+B?9R<$E`K`qec:$S(e6G)XUv-pǰ~6SmcO+gd2 D1P+#5FȠ-s{KW]m:7r~NH+ΊG y *4,&fSё}&sUy hK9*1N3pt,RA>H(%]Zk|֫ &Tf_Nu빁db-0B m]L<IwT[QREg˿1 V;]8`!{sk:. aPK\M7XگҌ|w%m3%eg.f"k R-__vtIК ~uL -uP┘wNXࣱ&taX>i)xk& di1Ґ6@*kzfuFi(7*~'W֯WbハE/V62>LZG;b˅(zRU[ ٲЮ<]l 9?ZMHWN&@% $`|C;K?9IL3`1jP|׃!R$'./\ -]wo>>ΐ^ȷ6Xs/`V(eLt& ΅eDnRq/eip($1 q`4 i$)}DᜐJIBIHBeILiLg&OlMOKsI'S+s6DnLPHin$)@BqbŔ[/!LeQHh 1 -5ʍK*e 24k8Z/㤪ů4 -6gMBB8<"r%,Cg Ʋ)2? |0^gv`&F@P:#?q Dv/6`sƱz~NJ4T.V(ӎ" g.r/x q\,9 )=RIz%2f(Xmh^dZ Vr((u9YE#~$JmiCP;^F}Ib*h0NϖsNKwtﺳL=xΧst떥57 `?ԷnХ`xgU)kCIV"Ns1M&'45pm9Pׅ.m5fFXQUh~9~1y]|_/7'ݕu9vR|7~w*͖N-x^9vuwt|Qp Ӱe zszsЛ >eZy*az$#+~j/"B%mdz9߃fnA1|= mѫOXRR<x" SHzu씥tQZR7~?xb3h^xT%+wIVn#>XmzH1|I~TtN9Z;S^^Ab9!w_OT{~t); ALRio&$`J!s1<'+S/QDZFxq%-;Bb] X(EUÎIwϕAY3\MX˖OHpͿxELǜ!|{ 9(N''uH/4Q#(:S~QܘB}ChNmwN_:%j<ڕ&| {ʢՙ ..{gh||29&lֶy k`xouC@3Y{l>Mf?KãE@ݗfw?LjktW-Zv !yns͂K/W/=ͯne;I uݧlT /vGr60~XdO{yۼ.>]l%-8si)xkA!7d:&+?{EiͿKH] =mT+o|^c8D84OC:MF3<,O.S8bl۫o/äi[ 9{Vnqi=Nx|ka%P lT,eg;BH8)yqZdń-b%WbOA=I 2@-[D@d6 O`tH}A.)Қ I !, o#R&g&!hRhNAQ{352c:YlkVmGbѝ#=3bOZ̕֐ZK/T5{=i 4`3ɖ'\謏-@rD?ɜCanv{g5|<ŒP#jWK~9C'NAacS%@vPj`УL` Tȵ/Z`?P d3IGJi;O郋p~*+G<ԣ$fi@s]hؓhΠc!Q4 @MYi&"胁bZ$YZ8 .t%>:hڠiSzagP;"ڳ^ UF3=_w%9=a~:w)cpA?hwCX8(cP)02Lo0c7Em^oo֘5]}|hc:'(hR"f=qjV:)}# NFZK8u4va(jXR,#v_i:DP#E SAߴC3cY{$}g3=9ϲ!]/O,hwV~TlSȽTrNScu!(Q"䏝! Z3F;B&Jq|-q9еYt_ףF%QiDםZZEoŎa12嵬]}wuɥSL 0^8cflLR].z!#٣sI Y/HJAœlzXS/T \ĵ"6 ނ g5"9T,Ax=WU$%sVCğQa&H@ig hbꍱFr2>MaK2 *_ai'{F俊v:|?`gY,F4jىs~VKeG˒m5)fU,V=tF2V},H$$2T&+, H3R[xgyb R!@\k"[du!&LzB]eYFK 'Iϩ1S9?J!/o{]˼CMޏw3#aP* |$Q)az2'gJ2F?U/ 8o-qVW1|K(ɅLk?ͧóӼp'D~XO#BS/ˑuolkaZysyFPz:N=@w˜0B֋aMWgW W Wzͪ>f%?^Nϛb87+~Ӡv>WfnuPl6;-Ǹ'[{b}Omn nfy& q8e`żDW7}/M3W7:d[ctiU9,&>7}YJl~MNHo4; XauCۨ5YcᬸsX߿ߞ8w}woa45)ؓII,p~VBw57ikTn&|{.]lob_/?; +o\G8U/ 옳\^6WyPRey ""D~xh /n~sDD/u}XlJQJ=۩p~08o7>;mcN]?k+<6"`пߑO@+u,F';K{!a>cbަ4x fi""E{n\>տ=bXlZ0鵷FkXHSP,q|%!2i\<$RgQJ+5N!T'PU^|[F:[ #f GWأς8ouZtvZ;Zw(8CNLxgq>CbNreF15,"c yʹ*w/^TV iO J̒Z]ɚ`rdY37!C;BsCHhWGBM ΔHkN E"*FhC;%s)#)`{cB{_o[ɒT.|tٷd}&Ds^fOݥ k{Š>@zz)FMj l?)Fև= +)굒s-ut5DKÁ\#RseЁq[D;>gry:}":^/8jz5\٬~X"ᗩ6*Fg@x4`(>q \p3΍\&AbA9$ՔN(!p$|!>Q),qD9"x d9vĝqelP `KV>^vx(@Q3d>Tan,~zKW-)6[szSt5DCKx 9F-SE#BQ+s΄z'Yg#9ϝry{عSfv&(dݠv{S!kE?~Ȑ=ːfsh:UOevW[j-5 A~S?˞77ˏ󸟯Ղ@_#N{q`~r@X#' %o8T*-R.--[mOd|+XGw:_?'zP@@RrQボhGdpaڑ& Jfpa04v:* -W.)M׎gNI6{ql7C:eJ=uRc skHd"tJDĂ)1Z &W{=!y$ᶐѭCAzʹ-{Q̉G2MLґjs!cJy XֽH,uxd:螎*c6k"QA@4&MR=Kd K(2,=Ȱ\!y&TyHO-1`A N]MԀ^ "DG)8Ȗ+H`kd[`l cpÈ FdJciJl2N4$qp"$8x p3.'7OW w]v{*ii NY{Y/gd6E̍Җ+JQ 6LMP 6 ōN9" !K@?L"qI{SDKZ"(vǓhZ2eޔ-r ٢SԺh{,kaX͎?@b0I?RUiD"EVE T2 hE{,jHdHa&$ M=Ȣ N%`jQ3̸!]bϝ]39۱'HhtKB1 I+~^T_xU3OX7G{M+?ꏋ" hF: l@6x0>Pr1FGs#V,_^)}ҭT?4eabƭY&-H<.^$A:r͎|øT3ǟaWjpwWj@.c/̗Ϡ6G|t5v8,J8KKe%YVAͳ[- {{aU|5+Xnff-&ro EP-u| ^\Ù1 c1W4¡-|zόU♙0TxZ1,^]l;y+FԃԮasm'*]ȸMM@ir$VLEҭ&"Ȇp5v)"4[.Kj4@fX>^@Q,:TV[ڼ]gw#/N2AM}C ھ1U4ui1{Te=qaăNPiiDtN z,Z@6$zsZŀrd(*'cގ 'Fܽn؟۱!Ȅ[r';T{3A{~gG]KFQ@/xEsREI8Yn9J(9dTCit>0Ƀ!%KL`ցY:*MT1`IHQD5Jq\9C8 F%@aրQʎ@vF==Bg%\~N'u2Iq3njZ4[hp=rfc tK6얦u?lҭ6"76;ԡ@h*QW'@)c&6Fg:n;sM.Ϊ-/ܼr ͺǚErŞM[nBܘaw|X4:L향9pa7Tt/!F3PG,Z܊nݜa usf*,E uхʮcsѠ2DBBcEe/*c.dzI$@,CJ9ᥥݫOYrA ~խKE:z9' 3%2ʕ, &9X4\'%Q[ wjh+ KG$0E2>-H+DK=;E`Zs"N)}dž΢;S9e=܂yT;dIR93KCb'E;ƠSV dm]/e+pz_ˬ6M?`04#9AY@a6P̜(@.0N *$MF&Jw!VT".8, wZ6Sz0S/i))a&օ$c cD*◓4r{F1cԧQ,N/j$f@Yi\Y;"gWof֯iuA} Cq袭E"Ц>Ŵ|3?ކYZ64f\oz~[oYo_kA;`楽E^NLz~7󛼜<^LӃ+ΈK>'=XdAx2<5^_=%ߤ3\^e>/#o0DdvMGzaw5QBߞ ajN?"h,N`> ^H%s4~u\Nsw  3ګ>MS6_?n٬ݽ?~+ 9h3Q#5"wm_!e #Q}ȇMHY,SbDHɱC$!)ixYmiTU,)/y/yA,wI✻ !U9ݦzZb=x>cOG$TJ^e5-, >W`ai) <1ΠQœ%%e ƞ>$s?} Q6F`Ce@mRu!٠9Da= {MhΜ~QPjQZa]J窹>N,m3% Oh'(iJF1<$f͍N6p/3J9hͬP@2@0S2OR@y+TbVE+= ?i6`v[J9ݫ)wsUb?15xls`e"[ &2ɒT6Ds%M)0͉YE\R2<ƨ| Q@"IF\h&6F  ƃjǹF6к)LvTgOsLI A4UەY}5%dr%ײ0w6Z+sQ+*9qua;n(,QD颲1لFK B 2o>A{wÜPqЭ ngz)%x4SUR TDD%YHQcPȸ;LX١uD8Ej\BCN/tZ-H$b(Ι(ba>+[Ӿm0J^ZaBrLGbOE!Y^sd'ʃssUeAz㟹D„?L~wx>]c]nלׅ3C0Ў -^8 >Go#jWis_AY 7ǯf?̰7.+#~,6qp%UvPF 儝ݐ %9ܟ0^ާYjrm$Z@J8j=!٘Rq5 ?ŇI?/m:?6 Gί΍dpe6Foߞo?AF0~> cvӹ?9)guý2/~xR5׀<ѝf_kw?W~/|w7|n 8P1@C> E8(HIaڸ s,PYB9x&STs9L6y ׿_= ^'-;H͵}mln)alW Fϕ? HCOZ0drdZ0dj%; JJ W؂APNH] OF]erAJ8vu'uԕDKURWHdU&̩L4Ǯ2իTW LR'2ZBj5!Ǯ2u ՕfBO ]erdUVcWWJ%zWhL]CgqAo6Cr@(*#hQqMgx7Q %Բ!qţ2&C*)@9K^>r/땉?}a2[|n苶E1o[o:Y]| ߼ގ,~%}>/\7Ҥ4i(MJҤ4i(MJ%Ҥa+MJICiP4& ICiP4& ICiP4& ICiP4& Ũ+MJҤ4i(MJ&oO)MJҤ4i(KҤ4i(H=:.$Sv2I]\v2>2R{RWRӒ 0h)SS)6^)Szʔ2L1_hK,=eJOS)=eJOS)=eJOS)=eJO#]5""3rU5Wi"uqf+I SS4"PG"OXK6sncL;Dh\@1*MZnTBJZ"Au:ԉrϭ"n3}#!,gq˨Qp20j<&*.FU'e@D*!8|pP"} ;瓧" &`Zl>[@ 't+modսeVo+e VMP‚7bZ ,qU$NzdtxSBPƹIh % <#:j>1e Oz0kWmVC4vn6^96D<[y.=]CmF4@uxm% ČzK07-HkNJEĎP' X!4]^d :p.vD\(¹Ԁ#K -HiFm`t 8( M(F}sCnn"?vCYVh1x߇ލ.c *i~0noI.0J[Pb0z)2 !6E#I8XlU"9P*FVG#BI.it*Sx- UĶ,YW/@oQj*#9vU\5QƮ(:vgYf kݲ yBr^?;\GÔpګG*+`+ Lr Xz,Z 99R)M&ŀHAv*P{ Czvt:}\gYP:qG0#ViqvgEϧ'OeRc.aU%LTeEn, r7" iEU["99XVVX?$"ZMnpju}\6ed"E0'Ku:?`سP<8L<ōݽvV'mmL "[mqY'O_QrǤ\EӭIšq$e4405I1p4Ad?^_Cd;[ά\'2$LdIn*to,dfyl DŽw2cD)سhOnXOdi*m֙6IA_E_byAQ(jXNv'-9NZ}fAq1脅h)ov3bKJ́Rjv~=^>o.ڈJO4j8+* *PIj*JTzJ/Gmtu)syl:pkY6JP %&u*kh]tJT@A!T nVկjweCLG jYyрNJKQƄX-#3Z2tT"Q$HgA-GpH7 Z5PЬUh߱e8I4^9S\N7 +[ ՋZ|<z !&Dμ.Wz\'RBr|uHht"W>THƱ`c)U1H*TM?|B]OhvW2Ύ =kvFu7 *;Mh%'@o#9RKpb&=&10L>&l)(, wZP@)T= )**YJHxD u8 {E4f4%D*@5 $( Xʄа-==h rI]XPZBiǤ^S"pP~[̽$[*ق3~n|4q⿣LCMvR.~g_4K>]Wf߯/W:0':A[_OaR/?-?2ːaPJb!]!#d)=r9ͥ8w'9:_|lBj0M,h,=FDݱ[}eʢ9oXp84Os>h6ۜ.wԑ8~ /7$d{\,ggbvr=j:A<,TzɱrÖd9vテ0#0DR*^UZ)_]H6hl)ixKW_mMCŬ4\NYuSOfCwS 6}+Cf7}~`s*E `<$ ͍N6p/3?IzgA M-x\By\`)To<.\MqAhsd(,7qj~pwN?&ܽ&=c`o8>{#?e24o>y7 <܅(Y+ŗg9o>z>/ܐO2/3C -Ofkv5/ cc*C)J-+B #v+fzî2}aWZϝ]e(v&ٕzaW\®ZM͹ J]aWo])X5~2 ZaWZ(Į`ejGpbWZy vvؕ!{2 [/+ͷ>]-Bx#Մi$s1cP)':EVL6@[thr0O&Wwg'Msd 1;w]OC]dg/|P;e L6/TB;*)+!.gi@"X DT_rh!3'3hF$]HobC3K\c`.}R), 6P>7QJf ; x /)IRdIS!ڨ2C.<2C.UTe2Ter\!yer\!FUOpMU\yer\!yer\!*} EOTfhd(WgJS4Ems~6/:jyE^->V,W< QEEMs]bum &m&4N3jsE/,@44xEIQߩ}'}xECtxk_ZLnz3_*{=ܬYȼݏKymOm&/`r\\$`CB5pa4ReZBlF8X씂"9P*FVG#BA.it*Sx-UvLճ^O'Yj}U<^:%?nS43;#LkN/{㺉"ǚ?3|뢨)WTVVVEWN dX3@r4rN]gi2)4=tkG( pP>Dʣ Sݓ Aznl $ rI@e$gSGSjs cJ*X ݺ'rjC1+\u{3ϯ'}u](*f*Ss:>x1W+:&#с *H6T6x9EYPE,:O^|R>]V$W_lO,8d4BZ?F̌7ܸsp1brs]p:dy{S_!!۽.-YhW j$fu#@W_ѥ}ny~ϑ`VxYB7gI7TA҇*&h`PL )q&eH%*G9K( l͇wû95Qwc2j`y LZb6!»[^j|>k4!օ.χ.wG(f!e](> Q0Q; M2Lˆ-S#~>2 ^B޽;.;\_s7]"֪ bASp\t-VFᚰ.v1u&onyR/8ɸӭ"3]h'Ez8 M>ZŁ3-.oG|o.;g9r9lcYV M˰u1mIv69L>x;krRD? j#֭yQ#moP;n?Tgw(Ϟ:3}CIsKuqun7qbA3U$YQYPim^[:Ry%FK^H^8W.q  :*MT1`Ii (Ebh|%ruPH}D9r5d>1q dgP8Moì׿C;=nnw;^j-=6l|scϬ.쎭Rx'Wn8n=ҝ> 2tՕ͊nS#OYR_zjYݡ~?qi>8-/~GCr7үb/Zܻa[yyW|*k!F ww,Z܊m~"\hl~ɾb9niSdq?7l8K*v5 ^t%Gt.(, wZP@s h2yS@AG0nsdB[-& T6Ds%M)0͉YE\R2ƨ| U@"tK\h&6F > ƃjǹFwh qngi;ٴ92V!0}=mAx[z[)s̖emG{n}RSԊJN`%qtDd* BH&4zXbPAsWp/T~t]ڗ v63%#SJAi47 AB !J.2(l'pO9B!^qlGX٩uD8Ej\B!W=8L$n1j]{oG* 6#M.[O15d8ly~ G#jHI-WUF0X{/ns;lMښ:],L l QL_G;LsrҼ)OxЕ^$G26YQ( h04;D. eƷթ~/y5 Vmy}̄̈́j&ddߕ]$co f.I?ֺ6RԳ3FgYQ(5 \v(5u:%L.5]^q&-TI5TM.Tţ~-u?Uo+_|}=n5.)Ja<敨弶 E@fR狉ͲM-q%ַtmlM(o 'wI>m&X9ͷ Nl{ 7qZGlu2dWHE dR@ȰPWʁ&~*Ď@oo}ヘ?w_뿼~滯a[N7t3.;p|C}5 5MIw7-}*vlevKvdj@M~u6^m=?icV<+% l~U U꯫R/W^RHMvEwl[uHaŤJ)+°9yl(OMc_f^zJ\lˤ$ ܜMCL=?RG73ǎp/eSJi4ɘF48ƕP6*#֑F>%zg% #­ZIq!P B&*+$@ylty@`I!CMI$ Lo4W(0r(ꙇx^oZtZRϲEq#}HQ.<N YJ1)ՊJy,T Gָ=o !Wt:n=xd&R"d3jT#@O(x&q S @Ii- u(c"?_V?]])OƓo@4oq_Ovdzyx<;T}Zn՚ov@ZOPƷ@Z@"$/Lت"|h>%77Kz8!.ЍM9ΤUJAYy%߮@IEKŨd,i4$Lˋ\{c_+hrREU* rMLٵ9\h4IrXbx :Wn)etHn?m(X|)ʀ Wzzb^Qm)J 3* xS/vōw:l1:O*4XTz󋖼Z)6)uSg[N$Za"z^`Zg2!bALPB E&>i^ʒzdb~vl)?C41CuOYz`+תǍaStԲ\ .P`%Uk (Biɱ7~t8Ysw!.-+;oxJf IytFGU 4gR&M2l[r7ߏ}f{:1|܇YCmO6Š 0*zǕ{#N^CG֧ӽ(M>kjݐ&_٬~RbӜ1Ck0" Ge(2B2I3 Q`DKYL [!sE/^*B> `4 FMg&;xR5 9rG SNy[voW/&aY𗾞Ւ|,Mot']L΢}>fz[fRqRIj)GW zSbЭ~D UR|H=E0/~q.{j`ZS ZJufR Nn3CՠռGU:y^}%8z&iK. E h`Qo$a$񓱟Q󓷓Y &RAI;yu#W^.P.hӠ鯮`R#o}= ZҌ^ Q1S$~F,F*S̙03Dx3^xlƄ" L`d>ǠaRaKumC=)L4UjUBj*0FU"LId9KLx'V!4ZqJi=ZzŠ:v -k$p Ll{޺πgfi >r!kL&K`4RbͼTtCW\ϬOSh:Mǡ`]s64:X;Wٽ$ұ;])jRO$NNV-C}tF`ly8RmE(i6ܸO:>O/Bx%PP1H%ҥ>RDz(#S0TT"k=HS;dn҆tiL!jY5]Uu*ZEPFJջ|JhFZ^$;#%kT! ~8H؏0JMv/V:?݌io}X]i`JP0_9⮈o9]XWmK#̻g*ޝv\B&ںgo,^=8(,w[x ]Zx*0' c38chn૳'f)Khƴ^}l;q+թm}A(iCE ݯ^ľA=')b$:n Z0(b\.ek}zQ}Gv*v(;[p#m`k@ءdQjFyF-TMދ;*25F"YUqo4lue~Rk k_NZQ9۾t 5P^*1oOl<(_9Ȅ;r/;n{gg홾s}G| kM ?X<\<@i`hjiJL$bW>\|W t:©yd̘0H괓 \z»iFx(zRx,`Ir{g("YmBz9%" ]t0Z@yW {NGJ6Yd C=ç|,U۠yߡZk:#(U7p'yYgmg];;m+ڵa]kUX );b}N7WcDC  VTS>\<m$Y/;N_ol; &3ؗ &>%$'qO5),Q0DRKUU] >(CF9`qK97M}57 >_?fA-w; 8ɧ2˞bk:8-9UMY*b5:HH%Ak6H;X)Rq?x<(pɃ0Yq*"򩈰R gb <00T kwU]e% PRj"SXC#w&@βS Rpc)t8kEոjWѼ'=@le`[ߪn@f_\-Ho:l(ozzJ jfPozFq&2S pSLTؕ!`{mU+Y¡8KoptCPFq%r[|߽~sBmFV=4A()#e%= i"<93{Ww*ddBB "n_7M:}2l~A ˓Dཋ5T_եׇګ^~0uDw{D%"Fq%|󷯧Ur(zc93ROG/(Gu{גA3-1bQLW -h2`-螎)cw*/ TFYdW<}t鞝G^瑷y\!y.&ddQC D0Eㄌiy&YfPthՎFx:E@'ZZFb<9*#ƛ\vu}cXՎo}ho'IyA b:z35' ˭'pĆ =~Xq]6I(7 bhgk^rpmo F6WBW`c,<0;aX?iP#d(xH%+ 褴eLjepUm:Q_,(Y)D<B]k T:Te"ڷMF,?ӍrGWcVЯY'Mw [^wCR2q/>RMF2CH:rh !F0rnnmQjeHZ)C}cƠSbTrKPH-[#nx;Jو,lme,+ eY,\{YL,2T7 ?nïM؜[حT<RU31((ѸT^3ʨm,xɮ[vZcDmpXRYﴠ 2y2S@A|T+% BxԠkˎ:{D<uIRQ5.){⨹sP+Q6mXÁ@iuNNi[Sj'v,)m^;̆aL4}-q`*R"av^bpun# +IJa\6X€9=0"ALjMN &eF2 ,Z#!Mhk19\ @i;GNRC"BpW7UorvTh#a dc@)&*a^gաu|$+hnQj2Nِ Ø q?+.ǽ76Yo9aG{oc52 ~ qqR|CWxg2i/ktVwt HQnsN9ò"Po|>'S6U)_Mi̕{IP&sP706")xVY-e7m2 Y^$O* fyۘE,pMiYC EfeU'0cz)P pJc+6Y $}k4rzk1x402ᭆST6Ds%M)0͉YE\R2Q1@rWBD:υfB)jcAak6V;EH0224C9 Wg|k2mR9)9n8Nhַ+ssɷs ȵnW.nVTr(*g wj-Q$Y@f B)& K B;x! 6>BgovEC2\JddJ P]i4 T!GDQ@;9p+/? *XLF~mYYG5^RY͕K(@#?H#тQ7jLLm[YY1ښ [d Vq~-d(GI4=*seZ>vqտfNie+J^_#Ty |]Q(18`K|E%J?*9~ߎ'o-I+yDgcN1 'JǛ 3r9ȱ?*pJIilt~n&Y<{&Q‘=MWJxK1ƟN}3\22E[r<JԇޥLUHZ~*o<;|^x>n 8\x?UŸp ޜMC`688BZ[:ij57cHc3 =2aa)p(|8=hxr[hl[;jc_58 Ě?rϿ҆6Fٕz)ݧapE/J\T/;T{._zv׿:ze? wL&HLo+0'Oyfᦕm5Ms w7j]E]nh:V6kThl_-7gߞ ‹auA$\U˘=F B$::$t%UJ1bI(EFh G|ɗ@qxW1[R0<;K0j0zx{c~t0SBY[uHĈ@+~n4֏ӯ?MWxN"v,p5x!q<B-'ASHAH'N))umsQmnyf:F[U<  Ja&]F.K".@#s6Nq2%Br5mɼښX*i3}Kv/{g'NW:oKf\Q$ r|i`J^^H_| :k֔z.hCLD&3&H|2Bdv]v6,GPݨ:g?ںʱsm6#ŵ!ڔBhWZP/8Im-eEH.(\Զ@&@8ej m-啿p@h`2 w<$G7"Y-BB$bY0WL&nu x~-!}\4Kg=74C&'5u'2-yXԸ9vhCHizO7C\j3g!Q0+ ^$#L%eQ6GQėșÿJ0wWJ ^u8c,wNǚnjgqGSo@/b;<Εib%%/&{}.&ꈷ" kHvWzޕ)ݒ쌂nU݄)* L,<22 pk3. cU2:eV-i#4N?r\٬nͮ`h*a0NLbOx&fLTA҇*&h'頙>g3),CqT9* Y@g-3qk35YG̙;0%꧍|</7ou=N/ֽuʶҙ6cjë-_l-_|WLm0njWLm?Mh +$XVy;!2vE\ej֟ d*罪;qÈ+idtlq% D9֋+D~Dq k!q+*vF\erq*S):Q\j쐸Br%7L*S 2#+C3N35 y^Tz?/?W!K\v#h#PxotF8^F;^oxpIm}LNeUAOSs!;kq8JD|0 <$ ׬U@ωfP(T]_{(c'{W5syqdX򐮲vjN &t( ـ Zښ@4IsXiyL x{I"$ɲKSgi-% -cEI_\]B]Kh\a 1 AB. AB.ё$tѠbz.bz.bz.bzOwLmٿ4)OoZ|}ӵ @ M. Z^8l!)a ?E#Ba=a=a_;Dh\@1*MjnTBJZ"AZu(G/䖭"7p}ᓇFvcx8[F5JeQ0Qq 4":)u6&R똷杢tڛ*Ҡ?p - >b/݀o;MF.}Bq{lڰUȆG%߲0yo}/?s~GIƅ#LP‚7`Z,q$_'=q`:#L]YJPJRڧ5&Jʉ ;LRZ𒒤wm% f9`XUE2Z?3C8$vv*P;>qBPza=76p9$IψXBǭk7]H[(3owvy3iKf}uvSy[&Tn6v$O ൲(W|1xOz]=c4~홹էˢjϽ_+ϵ̆6K eWGhryZ]%O>_jPMgWa0v ZнS/i4=X˓Pάάv}Y35߷eIv8\VQ_wVfIrI2sRm{/ iϧk[X"W9Xnhhۇi6[(«E3%xr.%sd sͳ%EYzxex 2W S Kk0;ٔ"Y&kɮ\c:Z13(m쎡Vx'j&κqIw`׈JZ+nmB]k]^ Om#8~|!޹n['#?^|;t22~{<:~EBHࡘMoG]*4A +NBX^kf3U1OJ>w[R /9Y%/:@ރDlLgdټ=n soA. o󗫸a(T h~_`ew?d`m\Jwx۫ͯ~L7k|N6Lޕy^!ѡ6N ϒ])ՔHJq`*wNFqGM?B <S\kŮȻ""RHxb 'A%ToS2Np0:z %&154rg*[:%* *7F+Bgr@\*m+Q1"M[׋맂e[Fy{X i!h-E=9,gӇwu-t6f9ǔ1ʋ4QA9$]`OqS\N0%@K6$e$8rNF䴈b K.'dFm!n,3:GQ; Tx㩎&ub\k\-y+`$S).OJ/n\  ǫ}&WhL~)k7,RvcՖF + j[8+D5i0==Qk &+[Z^XƁFihD18DQ)=X#ڽo?h.D]o/wq &mPػUSy8R~|ξ!c&H(mA1jhʴ2؄. J7T$52J1 ?L"uI{ST!K^":v,[77 9Y; ?^;[]my›}:qݯ=cyjqF`t rq.3 NcpRcW1R@Xn-d:ྋC L *|*$wJ `W]'R9Mi2)t=tk :AʇHyt͓;iXo$ܕ-v"oP+x֏sЙy 5걣B ⳾Q7]҂sGe4(8ޣuEE?+f͹h#~^Kb:휗*H}krs>/J_-x廒E 8@ #D,ABxwՓ`YzelԻHE$AKY\p^CQR#Z't(n u!c-7K!%%Ba9zrVd#%+M477q6Ecϣ?. As(,U+:6hl_nM( N})@.K:XNH(mƑ /{hmSֹ8WO^svTh#a dcӜ)&:a^kP.x&9sv>q:|[6@ۻ0GraTN܅}tV-C_(cJhmh\vp^}C)Bx!}DʼnKMx=}sdQ~v29i!"c9MǬ".)LǨ| Q@"tB\h&6F iYm\D ;nh 83r|~2GER]LIveme5T<Ŗ~Q]kYޡ] OYZQɉ@;n(,QD颲1لNK B2'_u2)yU\ϔL),No@xS- B\d P w8 +Cjq69r6F%ºWH"T{"He5W.Tk!'};L$n1jLLmg0iz-Ӵ WK+LuH H4$k.UTDyp.xzAX^t֣w5̄;on.Bm@{)B\9ЁBna]e &U :Tӈd,/8L9O]WSB y|B|yo52\.*(ɥ<.nťS/lY7z(go.͵Zh+yDgc>)'J r.xɧTX'蛅}{wR&*~;F|0x.}LximUN~xR׀m<݃f_pݿ+?T]5_g{c? i4וu] !UWvXQ $u&giOcH4ˌ3 aIp*fK> ]>9ζ=2}C-=k7\>Cߕ G :;k:tQ*'~?Bv/?}?w2 W`}LBIǟOO{Ojjo>5װ{jTO,y ]>r˼OMbʡ]?}n>,}>30?@TMGYxy;Ȏ ̿d^`^T#߂ ш^ݸz#"2*&) S |qF6ж1"<hͅW5X?ՏߐK+wxN"v,p5x!q=B-'ASHAH'N);uϧ<"Ar鵷FkxH ::L>T4.iFl2J)|ɔ\ &Tgwr3xqjqx/JzVԉW]쀷f\=tμڀ|[1B[_J1.?1*ea]$_q'AD/kdalH @G6ȋO 0[/'ڙB{s8kM'6DrO$с>ipT!PNggAڄ2)b/OwPxI!@fx_6hB`m\J w }۫ͯ~܂h2ݒ*^/r&)hz Zk{kW JnSc#c!-CC)!g["לKoL-W[/S c͎6|!$HP̦EIW$Y(ךY +zm29 ez/}&n_ܾS{G%nPuQéDxUhap-/;T"EP'pX"cNEc7wXgrg(A +HTȝ "tJT@A!T nV<gumqoqe3MۤRqO9Bf^\?,2PHPOIsW)ݢ֞>ލƬ02޵q$Bewԗ?dqpsq}@Ej5`^DR"%YLgOEʨ @5SI.gW(H2p.Vۚ=e܉P:@U.ِV,WAԶrV**j"ťsv0Hau0`DL8nirnf8^XƁFihD18DQ_-+) Bq>𱛇5s7G?6C9}bf=7u竲[09I.0J[Pb0z)2 !6 FybQ 䴳FFC4Rv'BI.it*SxJDC\(zr|3S[.+_ .6Ť6Y]Y,9U[aXO?rA"\^L |*$wJ ;뱳;HGS) є&b@נt^;P=ZFAP!R]`E'2΅HV$2A3⩣FksNSRXaY5t_W~vQW9tU~>D?X<ĤS"k3_Ldm&WY#k3t_ad-gT{Tkv0&7yQOp&:c[{4; ~X}FĽš(SP|&Mk/m՗k4g;]JJe~VџPQY|fqX-~lPWQ`&PbB p.* ¦UGb)X]ځB@M & S!@VpgEeBJd9+tW/˄eB8LFK^H^8W.q  :*MDpB\&ISQb1J&ڡ^18o3 Rh)(Gb[G9& cblT;ͯ!p|s{M7]MNuw}y[k XEt'tyCN ]V᭻]>]:W\KH4SW] 6\2[*1έIGMs}}}Z燜 G#x~4L~I^ud#D;4 ,6?m.t_@"C@g?X/QsEV㰿c.|}9/S3cepͺo'@wϹ$iJF &%0DŽ)&ep9~(ǏkGA7ʚ,mgȄQiE&H=I*USG-8w*@QVxECBr4wLw ~soPk.so.zzO-VoW \ X-[Z{;He_o>J^EJ4@ #D,A kl'Y 6BFFYFUSs&C p-qbkii p!T\*/IhmjB&$xK2+EfFð`\./=گPUIHeӂ(yTy(9L{^.$G{7*kKQ=l~[Ru2,c 85I*ʃ ༆(G\lbUN.`:'7Ys34B,lo?5 wme2%ٷOmtRe 5>[f}?|,s)SD&׊B}sv{oy{T?6uLon|w;^x!8s?Ld^ׅ"dϷ\݅'vo$چa:"|@! q8jt`żDcugD먌um>f Zc8<@ok]N &+:tQv*'~tNn|Ow߿!ỳ~pF>{wgΟpO&'ǿC׻Z ]jho64װyhTɧNq %e܇K|Ơ~a_r7 jVq?f_AQ̯&gC7˪PUo_Fh}՗Qhc9b QJ= 4fe8sѷ9{8gmcN]?k+ \xe_Z7ol}htWSʍ:% \F ^HOPˉ6DR(I4pvFssՈc,(͂XF"1J@j,-k~45Gu@t,arWyuЎ񗞎/ݢEwhXeg-z$U{ivW^v_nIMV֨s¸s@H,^o=}6ɒ3>k%q AZ'=75vR%KaW cx{崀ܓ3\ c}@_?0|饑D}_)ꤰMJ ̢JIf1 cEDl|xP Mw^`C^{` ]9ۀoH8:l]IF.X_hvnuMNj5W3+7\&62$eEvW>>F"$$/և tdϼ8vg*#~i-|Ij/XOAQ(IzXEdؠF ֵ+W}M3ovf4堠$SR\s$  ,J$,?hYL]ƱҺ_|YܾߣB_[>߮%pt6=i3科^W28@vJa ENΘ-bgOMα2'Wr*%$TIYKРd1RKh;|4翄"۬=Xǔ^: <!#d^a[B+ #f2ZHQ-d"c2>hb*^>< >%dIdYoGxiY~ - 3kY;QFū,|zK):\RZLN`WYٰY1)݆\WeBBR 1iźy(!D#ے#u yDQ%RHҥW;ck ynoF-ddʪC^[.s*=D6*ݿ>ƚ :i6 It=zK1']m͌VJ Y_]xyuK_/¯Sbd<&'kd&n2QF sКPS~ 2d<@ծqiJ!V ̪]d:;ޭExTCx~% j+R+&**Z*0\@q3S,M9ltQN8bs"@)"/TRO]B!xV>i]gPL7iOknkeogqxk=+pǖlz:C1gs3c{뛻t5]ҝkDn%mv$ZnBBwC>Lb:nxԱ{jɾsmݶ?0qb楑VoquȜq+9|Y;>\ziqxEq6;˧rKii%m͏;Ph/"hѾ+(13*Ԫmu_Jq>0tyy__ޢG(aSX!|{9oAbUpz @/՝^pIiKY(L^xEJG^!I#DV7O-x&9sv>q:eh8.*yŐfH:{\Jgd?8utv1zס󚠓)gY9zNwq< 䯗4,v~A56x$P>]j鸪Gߞku59[ucۧYGBG\@I+i=Bd"jTPE =bM.̵1T7JYlk]ؐ4,ǬerPQƌrchp!ZbfVunلSm'OUcuO&|]AjéVi/S j!fE9r|Š.>; Df43p:pTzrm*T 9R0؜4 +֚95vr]،3 .A<>.s I]3hmmM7/l@F믣+dTNhڡc o6i@,DdrVƍd@cA2/Մ)LֵnՊiɹ<݌;=lo*`+'D*e_=W( l y} .9kXPRۦE"9-=dFً-ľ$r0pd_JE'٩6n܌p/rͫb<،?Mch8hʊ5!Ev[F+!YTbQqM6Bf,NB JŨŮ}aYo6HKYKa*Y՝Y˶x!Hdk>fRU0%12Ǐ3؇|jd"u|HBIQ!!7?+t KzѴvuqGYRQ %B\H'#\1g$٭W=Wm@ Hv_0ÂQxjR=_j)EHyqyZg?㞝JmvØs⟣p=-{AY Jv@ZGBZ@Bs` \V?\1v.w>ѡ.#CDȅWcHyqO%*"l^T$/6>lFΞq|~22yTJbǦW~r6Y\ޯ[l)4T/9YKg6Ss_Ҁ ( +J@6iQhmG 0!s IK_tqyR* w׷{LʚP"Z,F[uL,9 dJE,IXy23<]]h.w4aU}_ MI5e468TXfwm_!e #}0wl [CS"#/xH^RltWU:?ɒ*'$ފh8ŊFLQaFzASs8wJ8_ /.Z u9sթƔ~ iwH4tNE.¨_ IXYRþq7fcMy}Iӊ>FEsB)|t@F"rGRP;5aizW N[lB\tH80xBi{qVRϓoV7|)Oή]oJaRRys6>*z6 OeҧԤ[ ɚrp>Rask?Uָx:f__o>/]V͂GaMWݍx\A,O7ݥNoBZlaK?ff4ZBLCW(4i'u;w@mۣe|a-oVU=@#5a7p!ȮJ'gz㋂{2t++&~W_\ /û??}<çsL;{u#0 .]G- Ap1 nD:nZrT^4UluJ&{] v5MvyAIBv}d J?~d~W`?icVN+o ԖӍOBoZLg]R(΅?۸ؑc5c,Ցnuc^-씕&aX<``|7+gexPQC{Nj^ o\!XYOEPqi,c aCw&z@ИEυڡEVan%JݮOGj\< T8VHcd -":~*XbMRJRmE6_B5r uP![HAMzM{}s<|սN'G0`ړ2r霧<3$/ :5 >8)=S"oO~'Mhãuopb Pf]/?߈ZPKmm̳`_ joJRr#gbl?aɛw <(Ar?6R»8*:o7x;dXAx7vXʆJMQrn+St!'~{7Aw8oֻhSL݌/ P5M_uڂغmlBjлmCj>3iGZ9ʖjv؄3-XhTr6(SX8S&@D}LgWcB !lcr3jwơio MޟșnAPƭ8쒎*Q6MIg({Lse2NJѠ߉Oׇhʟ@yjQLBZ)a!PoB\jV eMZY6Y  q6t\\%*g z=J1)@0#j)!.,_oFq%EhHvpԔJ䪣W@-Ur$#Z2BGhAS|*Qɳ2ŕf᩸r 4-ʿvJg_a+L89 &1)4>εғ֌1ʕ cxp`0:H8c!g^+by FHTߢ@?8i뺔*f2c\r}Ӂ60~SV⫺5d̛5Ex߆Ixn$f`l}\BeK*t[PQ^':fΎ1fϱsl=fN:҂pc96{͞c96{͞c96{͞c]՚:EXiMsl|c96{͞c96{͞co96{jR͞c96{:׳ >;<+'HT]/ @M$IPs(c02BSOc/2ɆGbxDd,Z(łk0:r)NvXy12|4XO2 l.$ 9(KlH~R[ïlm v`[z||d')DbnO"X}~׍i=~Ϫ'1ׄ"+GrKq\wx((.W`x D|,)Ɯn<;b*h}i.[7RU3~qӻH=ڡXw2imbU-avJ^SAV5&[xaYE;nICv\5XV,T٬M^Qaq6tp+[w2#EBHh-F"a4@d$ \mHl]m>n:,{0nwM]nwkA7CpoЌ>ڌBBj6:XS %ݙOƋ&c` jV0cy6ώ@ϵy*wl%vH*WSc(g":8Ap*Q*C0!Bʂ\ [X1c@K^ˈ)Hh>0vm5jDzEU "UA ؿ[}?z̽_Gc:C^Qg6z^^TYU?:+hV\)]LoD}؝J ^uwp4j k)|= z%oW_T~5Ջ~=WTMuGZ}Lbjiݯ.|U::w=[ ^;+4Y-{& o-&o. q '?@tW \(8L0+F~KE$"_hP 1 3Yo*JNғ]k)t}͇q+n,F> t=|fcSlvZӀ*3:6:$}eFP._@PLZ_ r/޽t\u$Z~o=26w>;x1zu6/m]L OԊȺ3\r[iT{TJ@t {P)Q1T(nq]\+ĭO""` ., '1bePS Lpa/jKǀLLr)7B($S1f4jR0#,RFXp46Y W @\C~,ޓv#2v??Oӹ2b jfQ;|z{zRI(m"Xy "DoC1ozD%U9+yyʹĈK/nƙBF]ҀijS:^0EE23&Ov-EZy䱴計4O+QyP<8ImЖ ل:At(Ql=N;N hoR0#4'(xAcyNXR.ku>ܟϼ:Oacvz6)~\ly3M,S^I֊3ټqUUuӼWwCo[(H@`A΋ FJmf{0X5ky[jpCn"xb) BBȘC`!7{&_H>R}iqhz(c#4oc|anFUu¥lM노m62 IkmH&@FK̀7Y 9X'ؗ J^$'GA-kɬSh[@&bMS|eVjO5`бB#޶r4@.ZyLjfzC1H{<9Id^^6V1)! &D`S L|$ά".)LǨ| Q@"= $y.4JQ#e gଶq.DzQ;nh ҧvdu'i`k OJo.orkz1Ky{+^˽ n!9OQ+*9qurpǭE6!]T6&aABf̡J:~P(,4"3%#SJAi4 TDDQ@x=Vkb?N*pC V5epTS텋 \CNpZ-H$b(ΙP QĖ۾oښbYW>fs: q*yboF]= &' V:(L';{}pO~8LpFk+H?=9<o[O.557k_~ge}3zt6.rEdի?Xd,gsٞHRl{[a .h9Ctd5~ܩOyTP|02=?tF[M4J7T?zU'0<H2u*L-cTd*%k=/ShhsBp9^=vjv*p2J2̴pJ)0sz2py*p*ST W/QMP,)9rr*pGo f*[zp e*Nw5'erzp䢅WBOD);U&R+Jh]/&|gyAϟNe3|7EYe 1 a3̥<} G6!Y.Mj,KT;b ^R$!L dXnI,47ߢTb̛WOk4S_~c?yIKr_Z1)X\?j5՜VLPAWCE. DԒ0% U`ɜ$2J!p&Ĥ˽Q%aZ!r*h`4,н y_@nLxBAnW3[IJŸ -Sm'l2O @,;hRK 2gA3S 4XP.)MhE3/ĥTL$ DT Jq&(H BQ1xΒj=+yY͙11 >-. մhv-oqE7_=wSUEI|"D7)W]+Edi-Jf`P(z~cX=-^`/oo!J7fZuzKVVMP‚7bZ ,qU$Nzfak՜UmY5*N zGsEo{/IXK-)5p!(\Xύ$jP IQ5˜S)x*#gbBǤ!nGZ9z{j}yrm\)+A oN1T}?~F`Qj~(QUÏ#M{0Gp9{9`k={7#h-fYt" k1{aRX4_Nz7-rۦ}}5B ]v-|3e^׭? X͍M^wy_9d^z(1 rQN!7yodQ@;+jv`?PuᖏVޱ~`}Exj&x3V.C.k% f+˺r 3 f/X&j^1`mwNJ7-u ?Bm 2pPc(wL(on6k؏#)Cv4W9I= h4AY}qN<:m={)\[g67_i ܵg{vیc5F6ѣz=횉q3a=ģy4Yg"Jh'5 4$z~Kj(ߌ5jO''|/@._'HJcb+N)oI,QG;uF&I<ZԷoԷ~6)tXu>ҳƕA Bzzf(3H*x:d~(\"hd* '+pVT,TT/3~'\_tաh)ntu)syl:pkDTNH$i*mR$v+1sDP[ F%]lrpk|cZ-j'MoUWt\v>om3-ҝkkC0fVJmv]xzUtI7\KH4SW]۬m|ԣ3$uSgw]c;|d߸î[#\o:ܝa8ozu]{M?Yn0o{ֳN׽OWۻl>Y->ols-_^q n>VBH{o;W%Yʠ^yEum-R%̶Of<n>%X'`xSȏF{-o)\І?{Ƒ 2r_o񮃓q}YCKĘ""e>?3$Eʤ$SC293TW]7{e@>\ҌT!f)$>Fu%fa~ rs6;ŅB3{Qwuބ+e[5 r-7uSwp :2clϘ>K1Ԝj-|A-fH{.e~u+ n Zn R8S4{K+U GP:ItVU" 9 "sUPR"D0ׁ.EWvpG F9Q2Z[ SxQ$ 9sY& qsȜ:@v}k}ʏ{#(CjC!ܯ қ"ʻAA=1Urr:yt)[St?^=u҃s[2pz0B:rG kc> }*GU6xq])y6 sCt%iH32Fg}|\bK{_ y[XjsXWt4K%8ki8n$F86QFSE亂CdZxҝiڣȁ\+,eL4#l!`21u$()I01!JַR$t$[vUq>xly6~>e}mX4iW>5 3ҸdYɰ20btETOJ9\F+MۯB5\.~fky6ˠf:6fޅI,AiGNQB" {!%Qi$:X 5fZ 0Ъ651EvU>+^>%"oJZK]P)llYlYOJCuA(puifr^/xP[#2*eDnSi0l1\\FGpK}Q[X ֦, у8]c974sgRmKȹ[2ގRv" [[BhYB' ^omqrYNdon38NxV~@ՠ GA L`uʂ%rxMRZ02J$w %,7 [Xh(O!R!hHF81GbBg"Ec۶E;iA8r#F ʹqp 2EUf!M4Kr[,s1(i41JjK1+m| Vq !ҢDC&.I)lIFRVER[#nyX+bcE0Dlm}l*uQwb)H+蹒}!NGIbkY!9 Z@H"{Θi7+ҍQ.ke2 r3֨H4iFz!\ymEi^uJ6eh:[WZLR*OZ%&P*h^^ʝ\܆\yTږLnAXg VQ6rGY%93\tҧbAH AJQXH&+R=3c7mʾJAߞ3>"oN&#kiMB[w(䃸ml&G& 9Lى TJ|0 QPF6jIryWydn F0E_*l!`k$`"ŊhU *uzf{<|Ü6o}FϞ]l^#F Y^{>`}zPȵR;syR_L3R{Vz#I՚MpsO?CRߟ iGpF&OpR}HN+*2JE ӕ)"q4(Lvm5--6qQ!i!Kk9 B9dܗxh@ЧK$SNNnH/nYw-D\UJܣvׇr-%`o0*5Nq؛s_POXKGik)Hfn~kBD^vwd{ >Ŷ2ׂn]pFf);R n iN~޼P)_+N& eBL *0V\th%Qѽ "AKE/7ip+eYgѩC0LHBj% Č!x[L3kȶ`Ύɹ6eW#]nӟt tu[Pw"wn::-F±}WԹ:!V\2Zjr> ZHy`J+\Uc]9 㨮T>{TEElI (93/ .b j Ʃn;n-юi#{ΠK|dIk جVP+ᧉ~IJj: & N: pL SI4ȝr̀󬤷BVRǔ`fb+'a(iZ RW(Zfȹk8Eu_3XCwx߯~ ݨf=ͺw[j-,z?]wHmHq0._܅f{UO|*\A/)l>7F[SP@52c<CvޙiRMmPM.A&x4YUa˥iʪ#(Ww :곎(:.mNS^ɭRd#=]o~Nߦ3Ƙ€)PJ?|^NYf݀#糬2B)^` "sUPUYR#HA,fyn?O2i~hVQT YQPEkWxG՝5uָY:k\gq5uָY:k[4g޵4$뿂J /awJ3:ˎZyOq,` eHc*gF#" y65Ϧ p!(\Xύ$!\V&A'YgSGH,sNTU͐hU HHki3!HFRk0V}qtNTuQ`>>@$3sK9{H@e DǙmpa _E#BAGk j H;DhhzYc8U8Ž >!V ^~EP'NƟ <%gø)έr; ?1L>}wZ-F%2LQT:DBt7[Dx-ԟ͟/z'Zlj{Gmod0̾yNJ‹ mrz?d]摎0:zRxFLQSBLx`O.S!ǽ:{+fg`;¹~fCfܻ~(˶+SH&XOőL4Bj5e^q$SjőXq`|q#6Vۓn+Q=S8c*/T6F-0iY$kP0;@U.ِp8!c6AX&YfPuhՎFx)` %8O` y*|Q-gz aL/YtySYiz߆Bn}3AL5d^1| UbWQ{hUk4ذH\]HqkU[al cԀ#2|eqKs46eh&OC[/~󂈜\j2<'cw9}qZzl'D잭Jp۹zt!EL6EFi 1TC FS/P!&4HW`g,`oHN;kd4Jc,Ek,~)DLhB:D4-,,7 Ǔxz|Ur cStZ{,|RX!.R;\֣aJ8U# xU5&SYj=V@GS) є&b@נt^;P=ZFAP!R]`5H&:j枆~#c\k M>xmy~WwG("kWq9%jPaB rxĭ,:l qh`E!Kć<+Wj'fחC3^ ?V$9|=;;y L}=-KkAw]Բ`4UoZܷi S.oU`tY*7)7;+U۫ ~fG-0z$h뇙W6 G5W|D2`Vܟ`XUB{lςU╅1Te/ߘzXzݫĶMD=iK- n3xc^'Yh-Loz8^hڜ47Z܊ jv^cM};vWi2$iqEtIs7QaI 6#4q$ѠqQ5 4k3{509*n!87k&Zl !;ef(N*z.zGM+EQm=q؝r}`4~xf4'HJcb+NPSޒ Y(wrՠHẂ>5W/U{~޸W{p S %!ؠurRG%NH$i*]U6F)D;+1sD-]lrpkP(daZ-jgMtԛNnƃ>5wI.suُ?[qܻŀ\37uc7A|0bzմ FKэBZfGf IH)c gw]c;{j>po\a-n.qܷyK􏷸hd[x63=ay%{;\z\a۳ۖoBzc# jJ JP\Xd p4^kfE4 lu8ޭ3^йZACZ5  4q0|1AxGU:ji]7Нtexp]RNjsYf{v.loR]8ՂsG4yP5yCCBCӊhMUD[:W둪Wlk?Qls'׾@b\LG jYyрNJKQƄ\-#3E1"ep`DHG4ׂW!Z橏Pn-?҉"(FfّLj!AgR.+GU譨V-d^}!d.t * NѠB$`4ES-PՌR3;G 8ezϝUHƱ`c)U1H*TMJkbl֌Q^ta1x.º.y]uvd.n޴~şpe8GWN@m$$' 5`C> 59¶ĉU"ny!60cPmr'!(fP_GI l."95’qy(Zw{Z\2`"PyPijM\yb\y)׆ \JZzR@+P%Q<#(|úD lv٬;~hb\( kDY5bՈkXBĨddciTp>cN`=G5rN%ڬLtJkXUT#"!HhG$AKY8G[FIXВfUc#|[W!dWᐋfkmW_ shRWVJkМT _=i&oz͋Ƥ/֥6`]ZE?IZ8rjKۣ3_}xՖqm$ aR sLx47:Eb@:#&9rv>p}8oYKf&̞L#b{HG:+ wI;fl m jͼGSǛ(_j|͛ߘ7jj?ɑ/oiiID#{fVFrfVP бAs4 m ЀVzvHUQz Ϥo~TyqwfeOx=SdxVNp29P>`4!JR`PQfqI!$$?JX( B\h&6F iYm\D (.(CS0$)FΖ^[>?Y꥚opkŕq`/.oW)LΗȭz{#k|o&+EQyYZKGIPب"Dd% CH/:uW*46*q+2=v?/y]>5 U)]ana]eytQO"j?Y)ߔ>iO3_.+) < hyo'+e6gITQB9aW\I)mޜSTqho-IܕqzB1WLo 7 s?fNi&Rś+)*~k>^lzyuVc 5=[܃-~XSfMǚ?qɖwPP{'۳u]bƇep1c/( un]Ef_Å.0Bf;g웆OcH4ˌ+ aqp*|<,sp9{YGged웵Y5[z6ą|nQw?d<;(XaݠˏpP<?@vӿ~ӏ?|?~#eㇿ~O?'\i8)IXO'ᗭVB[L5lZ+U y^EGVv;og 翝rQ۠9fzӝCA]Y]IX|F1'~>DGW2F} 2t*D#~x O㞒>DU}e(ՌS:nЬ ' 0.Ǿ?'QGODk.2/m/~oJܨ1:YeoS@M!!2DC8q h5^;ޕ,,_axᾪ!kHmx?o"ky|DItߝ b X^\]],dwW:Y9uU:3/LGBT (6, BaMb(2/$ T9ǯ-%N&NQ?5^Ȭ;e"|{W:W+7]^}iߨn,bUVqnUC}{*%[V鴶/XٱEU׹K+֋o[T}sݏW^'WU`WU\ԗW,gWUJ-zpJ;b./x Rp g~G9'yo<[MN+睚tq-:X#+Lc+ d Q%&>'RrՖI' m$0 p Р,(e1RKEۘ\DP2pZV$!:†V2H7g?Yսeڱ)ͺ&˃$P _/t󦙣Vٟm4J6lnm׵k1gjp'K/ӋvOMVBd]Њ:+rSFz"ձ}%[o1%yJ D--7iLJ"0+ cL|{4lw-% oNd߀qW |WA8Aꫥt4Vcս^x@˛qF Cnp!h;DՉEaFWvLX"b-JJ?`qH$C0JQ 'ޒPH9bLQ]X:0Q(DYXcSv\3qq(+F2﷘vՙ!\siq*_{KDP xͻZ|G4 h'cTf2Nb%= c($RQ :J $PKLhR(RI @ ^Iե ) 2[ dP7ǙfO)^zB31um|=S$W{$*Mc Y:ObEE5EFk@*P*H`JQ| $)k ѰTE|$&V5fly;ㇻqYWpW!qM11*7ey"eORj?7Q9|t]v:r9t:8#z3@ 2k-A,XrbE WX#[FHB)KCR[Bwb?Vµrè~czyzn;)i(Xtw~;0-"6$yB* ԗ{dӘs#zC+gF6~36l3m| +A7QD~,a|!Qg%בw ;\['zGeC~yp~=0· W'C3ƽ@ϗ&ldm¼*dbm\ *zYR _Sj^AVfMN`-YyX޾kAO3 \#_f_pibz\k +fS˺5zĊ+őPi͒ujԭtPR[曰]@X+l۶ӳAO]=ʝʭH ALB>e& E{HOHbD$2+KSr) WrN)R(T̖"5 @ |\P1wm&ΎASl~;?n١+=R=?Oh,]a-Gc y˜ζ-SwᝧOUHcQKJS%BOC,Kx:n}}]vm_[3/ܼj)ͯnwknX/%MeY%w.u5,-χnzˏ&}*1*1MaRxirmmrեt[aiQs"K}mEVJ/Cnm+[b` B7~32UZ.zdT@c+)t-a%Ьy˛JF4K(e KQ`d!;PS t M.eEWǼs.Jv/yKmW߮կA!_ *?MW߯j*]đ_CٱOxf):{]n|3 }@3}lҳea]겹q+-YAQ* Gݔxv`%}-:q .K(D]0b"9Tk岫t <$>Gmb"2 YLPkR.YoYkW;`\+ߓP ~L^ ӛ=ESQ=/ͩt܌?"uavLʏmcRL`~uj"8Biʞh0^,ߺ$ht1Iա )gc gY?;g^,ث%LIɣ*K:gLZQJ\.TIv.*$ }>I|5}NL?(Z(i *pF YD#IH TA;|qe /HJVfybe ()^F֟"T1C`%qliA6sƊL튚Ԡii~XM:ɾ Tg.qT)*)hL1WhU)&/n*#c fOK1IDMzWC5Bz'zReVd 0HFflGv\6[b!4B5H*3^SgzqQFt@G)cG%XtKF0Q NG j$AkmE r&7nGu{!(&B ںh!XHN!vcQ ":ߺu+q#v\P\=jzN!QrEiڣ4OhgYQ_b23C lEGcMb[Sh82FR1ID6[oflÉQ?/Y\mcD="xdKF$3I JE0z.-۬0=dHµ P ?+PA'kAʁ)f2&ݨA"ZE2yOVZgI,s͝-굗b(%!h$]ɔZ L5,k4M%e)`$Z㗑—󙯠d6ބ9 &BV&|Wa`nZp >w bݛ t2٨sO0մ|dx ]]]Ǭ|yVc}'ZcRwD)+TB8 لu \bVa"5`6^$,Io-쮒ܰaxOkr1wkrWlh(R/׼oS5,ٻ6dW] U`\8e 6aԒTbe$Y#$ـeqHvuMթs{V$)R:M‰ M:#l.Dj4Pkru0·' ~DMw;h[r*HlWdל&`(/2F$B8{'|lӺڜW/ûK'M_ JK)Uc]ڥ{IOM)}J6X> 4?߫4I;v-Hܤ|A^&K.;\KWmMR=$tg↺ֈ(W賷CXfiK5v{Aj v#GdrXV-ʩ7Pt<ǣ"or4;^P>@_z, 8ma*xۑq<)K^%Wj/uxETjokUqC/'7XIQ$½l ~zdyimGp\uo%{ክtpEg\+67ZyvVz?\]D^k yJnޚ߳sj{ $+J,Y}\Ðr\ ę<'u |n;tnNjl~vY5g&G b[IMNU&meԬf%ﻕ[_ק˵I ׌F|:.Xze~7]ǫl'R?l5iY -TߒƉ"iU|n\"i#)(ad2[KrAUpYk@mSŔE+%i|Gf-Irg.43M\q\c m)D.I2d98DZ˧ v8C_p!h󩁜gGG'SPXj5R gPɒtL*MB-p 3XB266D4c667:eZI&G-bg0ᚈfteYZH[5!# mF4**RídM| zrōFV5ZE,l) ~%|m \B :K9 "[QØ. |ReZCB'dHOɄlO~qZn&HΪ,J̱6G$y'R,ۢE ׂLQ 0:#I$5x7[(c&(G(( E?J!u iY>k%)hHcx RcOJ=h]:g(.zCD v%d)1j frYD+@]A=+l %DvψXIJEu)I6HDt Wh\ XOXjHfBxZSنյqxR& &jd-J42苆˓ȭ&t9D(h0iQ+ȡ( Ѝw*# Xʈ'WwAjAt&r3Gz7 ~nVOQBWBJĊ&2!mLR xaHcXR{He&4I坛QAޠ|ͩ~W0c[o! j ȂA MHC36Ӭ12{JЭ))59 &܄>TK1UM5.J vFƛb]JAʂA[*MZlYBPM2L$%Jik,T BA=kxD)$h~C?>8#NM uJSlB)vN^"{]߃. Pݗ5 nlK`OApNU duZ4K xrHc#hV6 ogiX;29xZoVMOd3u nm֋ q)8m|RV WoqGTE%AYZ{[(2Xp;$⥫s2YЩ ֏D>^ՀXEޑF9¶BM tYA";o mxGyQ'sOY3'XR C4+Qw`R;ЋAuD! p9X.h 7%x;a J)e`WPJZ%(b5 I(sǂ@AK^BP!)v5L|A9#EK&V Ap;=$ʀGdص7F!t Y@i9DP?z Z{hw Ed, WW 4$1!ddD(!cȡh lͥV2yL+Hr8%Eom2&iS#<rPZTHc8nHJxKTlTbHz(Gȍ ^I*c:R CB&H r caab4 i5`+BҔ*cI\(\B =:]FV2 -ML+"h\mŗh ='!֖Mټ}-w)۔w)! c}z?{8rbu 0Iv1@;̼d3VZQ[0}OQ,ٺڔ%Z硻b@s {;H +/RI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $, R# ֌R͠u_Hb0N+CI npI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $. -zE4ToH V@JiDHA{! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! tA$Г靤췫wRr]PwH)ֳuWa|?Qs'p aU\k9?{p KU}]*pE:2qX) tWF=rW`%{㮊]ko VW,qW,퓻foU1f?[VгqWtϪጟ>aS'9\NOȤYꏃ;_g1Lm7~}UZ}Ybe\ 0'-^TɆk.*sQ2'QNT\PFJoSN%\c-+O5[m38Ϊʺo˭SDí &)JӗXkG4`%%8^|MX_u-K鍻* xqx%+E(U}rW`0tqWkX:+M)G ?ꪘ^䮊\*V TW D]޸b썺*{mUriAqWlק`z㮊7 =:rwU%Uz&d_㻫/?|6vʓQt&lR'v(M_}.Rg*9+]\ ",$"j_fmy 9Gxvl7]t;V>t'6}vkY/ ͒&)4z֜A5GȼE7jjOG#_?FOmru[MoNOXG8P{]kMūegp*u E%yYJsc a^r?GY7R;>p?_[blymh22QMU*ޖْ^>Pvy oBMUrj^oEw1t9\ ]n%K5,,RS0f::5Yk>aVd8[8R.^q0S&w=4zp7-rCx I%#yR҉`UL"*/,!bIx!$!EJ&h 6)UJ>جs{%P54 b>B煣oiW5̻͸ƕ#U #IE`l by8{+ܤdcY^,I?R#)Ha5u^=Eσ msfm)xߖ=2ĉkޑ<;[_]Yp*[$R9p`:=H93!DPrF&#R$>Y/UJ%4͎Yp&@h -bNYH玧;3g4pb|l~Vnvvc}nzKlTf&fE SehMt1GkeMts_M9'KYM0օx-сU v8BX"ɸ;)M<n1>ӨyGheQ\v0"+&,\hRf!ьBؤi4f."eNeA 93w50- J*,Lq% G1z2Dw;AwCW+ďEp՛Z 澶+Oh%$֊%>I_ eD(6/SlJynbed.{Rd<q#ݞ>#LJ"I$IK&2|.x/>kAwZtt1k9in<=Áߞu!g^nA3o fЦ敿h_Oz"yD2N ]x!晆O1a^T`E+! G-:8!TJ!)d9HX:6!CT$=u $w/0<\fCx8ZGjfw_?m`cO&` Nzn}xLD |t5ef&sn|O6zP,1y3Lo"=[<֪,#_@S 'I"Hhg!D| ڥ2Kgty^waXG9I=q'8\W_.8;cxUqʄh!: 9d0))6̔ a9DT &ԉ&6ۛht ҝjlһXWݹ"v|4{0\4Y`)]Jw8ճƧda]>yPm>ǣѯt;N&A{?|]7Rh۟ݝpak}8\Wt~,,pK槉R#.BF7l^u.Bv#9DT!X:eV^јx;0dT,Bv{F!=]9cQ$띱,KC)q.@|૘NY;ʵ:*0d:d)ϋ䏱eG__Ƭ1!%*[&d$IQMrzkU]Ũ Ps{6ĝ`%Na'@g#9@&KpbvD#Mҥ/EvFð`b.=_ ͧxDB AB4rq\*L:r-w+~KB*BFdHhϣB]ЖDdKBx4 #g>lYDVǢǮQֈӈFڠ%F- #K \s29s*f`gF,YJH@$m(S<(`)}( Z g(5b1rkڡ^gKG,%EUX/N/vzqڄd5%)rZ"@ $r+/N/C/>,C]XV0YlQa'~dL Lܟ =P!dX*ƯUoМԅ=貆r >F;AZR0IC21inK/3PWkgGA-kl D& ѹ Ӧɞ<َQ Ў583-h= Ŵɫ?/ol:PpǟC N#w@q| r_~qԚ=S;RQqcv EC$)˽\9k%rݫk苗Sx d&td]VIhto;]Ly>,`<^#.Cnh{=އ)jp+PF5~ȹ|NkD 9`EIؿ4t^Vӑ"WhHVy\k.j(U'HiRsƄwP`O:#7@su{\ٛ=w)ZtyW"{E7P#j_Swֿˍ@Wǿ~w.++ g4_E'GA]uvf KO' ;"'JrN?u/I=AExyjגD ]G'$:!Byt\M.N0>Y48B,lo?4 ޕmI;J8f5O1hO'ߣ v7{2'2Ek\+.r>J5\{yW6_W4\?^g+7{i?i?+77(B6t=i6 __|`q Yד=Xnn Y"E! q8jt`żOo_'^Έ2Kuju3 d>?FݕzyhazK5]|ON/|/߽>۳ۛo{F>{go|_p&p)IQ,p~ӽVBu5lZ+Uޯ+R61h~6?~; kE?ۛ͢G~bUϠ+a-H6to_|ܦy3 24*D#~  O"7HHa--1[R(<;C2J/lucg'"~ֹaHĈ@+Y?ŏߑM+7xN"v,p5x!q<B-'ASHAH'N)usՈcُr鵷FkxH :Lm\4Dd@#s6Nq2%mBل:Jxb\?R2oLw݊:T[/EW`.|7}A4lYv< `0.|kMm Z.]l;DHKJXLj(f?TBƃb|%$@[ʊci]Qm LQq:Um ?P.lw<$G7"Y-BB$bY0WL&Rɾ`sB:"AP"lCYg/xloO|?d8|LVd4gEXvq#*7z?\{_nlJ jr> #J@+HTT!93#;wWZr^JNң(9Q d`d2Q҃3a̸$XNQ1>XY;'&IOpnOKZsj=|%?<,(Dh$Fц @Dg!8 h#RxT4 n5cSKH!oMf"P $Τ e -GD4SY@g9ۅx }œcSJ\ˍ ߯~^f?g-dzG/,z?]v5T)y;~ >;c̱MQji`4j NqEMZnTBJZ"A:ԉr\sr}MC#,X2iT"0E3ahDQuR&l> HE"6 x넷K݉?mߦ1͹%Cv7 ~|=*7Ez78goPƓq:h Ny K[.rBU#{@^uZjVʬZmrx?|^UR}$$8`0전odŴ8:,b TCe5sj9СdCe#D\B%M! l>j+ >1(-$H4Rr9A5rIqR,ZҴˊGV\(| ns|thnsLJ5NsءCt;K`õN(y{뢳 ̱`hb,ta/!rƝb(ZHS )L"BV9 8S?蓣.GLiY+mSF$/j.ۉ[Fw`͏iPW9|` pƇCPҌBKNkpq6P#%;SFƓЩ<_1F&@yAO)I"$JB> Ǽ蕍yCNwq#緪bv*8}b^kˤ #$pC6N$:8 kvԣhc;/&7N2X#%CCۈz(*mvaY/PYFwzƣ&DcFDh8 b]{ˈhϙT˞Io(-O]n^zL5S!vҭ?ՓxWԶIg ] 6P.򛟎͆g.ߛjߊjT63I~د\Xx}Mkrؠ ._)5h !$PIcB8Ƚ$4YP4NQz}:8zoI_vp?fF'db]E}yo [U=; .2x>W{|?a:# VqeXxl:$ ;PO#ԐTdyI$%6%j#f׭שs AP-pgEiB J0)PMEՅ i%^8W.q  9!X4 Vp$M2[Q4q%ruPHr`T&g G9&[fBD4bkbg=g;/cr<뵆p͎vK_m1O׼3S2ճUxW/&;;6][B>ҕ6{%gz|b>ePfן.[<51m׮j;oyٛ\=F?ni_ӗy4;]uV, Yg~ڳڛu E.hStϤ眙jgCϞ {&<* ˑ$x('~j#_Pz"+PTK5HNo++6iR}8[Ei"#҆>8jZS 1PKwQO$A=xIˀ'U!2kԇ4Ӧ󏥣 }]۽i\4*@w34 >׿+sX56rsxS}l04dG-5f7?s8cʼ'"[R|2c@Z}a]Jr)(P* BqFf:tH)O™lTDSai)<1 y`NEqoS5eWVx0:z.|BIDN` ܙ9"N H\gҙ졉K *lHˤ)IpH圌L{i5@Y]4Nݠ&Yf:GQ; Tx㩎&}g\k\Ė<҂F0e$ V߀yb[Peۑ۫IXȖ5Mr|C/_;}Ces_cFb:L| ;ĶM[ק, cyVgoXH\ QY!ʨF]Gc;=Yjpj2ḥ9TuQ,z#( M(F(ڮbi~iI [R6]6}(t ; z" !c&H(mA1Åt1L+CM8Xlu{CErY#Tc)j`Q("Rg:EK)%"Ml35q{y70vjwS!W ߯SԆzEY>f \)ۼe¦ᥟG[ \8~A_vJ﷓VJ;JʥTqLR =.<%q*WkwS /T!x!/#⢋Fw >5?[y,)ލ.6L^KկTfE;Tϳf5[gh05PKpQS*-ϠGcq=EE^;2x^CPۯ_:o#R;F(+ *% 1 dԩOkV٣/_j, C%)#": bYD3T&Tv'UϞxJ|zƇ o,NHLG jYy1#SZ2&2b8SND|tT HG4ςOS$Q @}5qOxըrZtZ' Z7Rڋ+UT A{!r"miHvS%IɮJR8e=gk0q,؈*U1H*Ev6!H*aak#X-c!tX04̌"773!_?P~x=U",PFAI,*D)9ؐ]C p-qbU'@g#9DaK%8 A1x&Ȅbtˈݚ8 aAnmtz_eDD΃"h+N6T`u̎ $NЮ KRYﴠ22C E{bM@]'^"qD օ$TGe6aiMxX)[ED2";D\%AM,Fo, .pg8ϩDNi Kn>$u(S<(`)e( j ϑI.kKxuxu:%E2.;\\9 G6!Y.sAejG,AH(cB8I K]a[x[G Ϋ:cg2֖ άg}s'߆o8 '+zUj4~uAsf5Ooeo﷍RqqA wdy|wܪdߢf9K νӁPvUh8HKU60)yH9&voۮ&v*30j?l9ײvAVP@Az%h1ffw^t)ٽzMaGS ?hֹw2[ *X]! iB64 *⒒g "B31RH࠰4Nj"JW˼C톦`Hh5q6ԝ|o(i;KCp4MA\7Ky+RU[ '7t%.WnrjVTr(<g wj-Q$YF!d*=,1HHm9$XX _-@P eLȔ\aBll%BWg'uX1u#C0ծ C]˻"'qR C壈hR9MgOϏbߔ{$䍣+^(OLkl?g'y$$O+\BQQ,BA텾mp6^ M9y~ԲcY\ݳ.ٶ]l;"\if`!U1UVCb>"\reɹԔ<,dzvEDb_W9ǁ+eǴvU VhઘU87 'I6(=dr rCcndbn+|!o?CpH䡱Lۆ^lќ CS.Rg7snʸ俍c7B{] xFmC%8Zi(OƲ`p.P 8}p8y*2ϖ٘yLD*ŅG<&IOuT,dc61ffb!n i#}fB=jW*TWW4j:tI8˱w4[|Q ֢>tlO5hN9 ̝h UCFr9V=YBs~-<>X4ɶ[*YTmlMԐMȃQE4!K D M{@[4 "C+*4Q,,p12Cp,%g2.>w>&]V3g3l^N~ C?.kSWw^6ڊ[Tx/b,: # Uf᭫]]uRlF5m2߰}\BW]jf' /VSt*PKfZz}ԜɆJˬ5 dC%n_}RC͕w [GΧ1d\cYAz8)ok9 B9d,)f$SN$>I(|a^Lg%ilӴm;y mvsf R O'TWLNNMyvk#;@7!ق0ѲHyƘHxZ01U2_A@Fe3.{:lUt[tȬ*yvJ0LZy;ƻ-=Ϣ1ǰ]3;>)?~.C >)fIL/F^si>s#MU#6YhoWAr =ilOc'KN*T[VpD 8j3@dFp!Ǫ4V,i F"Ԡ2`܆>dAQ8-<V"i"d/sil5s6Wяw{fMgnb۹Z/kE|Zf5)nU߼^$҃s[2pz0B@uI,Z8|& }XuGU6xq.x~,Dfܐ*jI*4Gp5s6<r$8phW/!kX)V^YJ;!i8nb%&p55u| $ F{ث^=z49pkxmy9 ~nmH>zFSnsCN1JeS٧Ƨ gi*K'l*uhdE'CfHȞHQfga`xD,cSm &d#qLAe:o[ock_l| .Xr7i]>4/Y}豄 7~A9SBiCP</ѳ(HNeC{AҗnC o 1&tSF˨9!&X !i41i1}}ګ:s!#luJo&1G_Ku+#Un:xiH|&"hXql҆LO.",G%쁧穇#{CtVI*/D9',CoDP1 0eCSy ]qPGGzzgoG?BdIƤusTBhSIxmPPF=up~W _1XŊ1-A dԳHQmew68 R /+ etx0^Y9!HPrs),0#a?#}^ }b!z<.pӬ7w]f|^\Old{&yG >KWm'uA]=eߙY@["}eT&m7xdMF ݖyfU<%stQrW~ٵl*_O<ۜK Tf\$IZř!j u:1Hzo"hǞ'ቓ״o_^ .{pQEpE@{k or`QJlE!#JP=D" L%ЀT%$H(4yd*7n5 z2+C^?XUQy odJE[ƔvZL)y13 'R*;Y^Dp_&ӝC`%6eAƒH[eZzbW3TmdffdUaa+Be, Isq0˜ӋnbuI-/oi[dyUZ`SrYdB/IQ FF >,)eɼQX9tȆ"|L lJm!/H1r9Ng2EckefazDZAl"d  "+Uf!M4Kr[,bPh(n1XŁ21CDCXHK6j"Q)"j4vjlVϼ, 0 "V"G׮b)Ht蹒}!NGI0@kҬ]!9cuƴƨ@r2LH֨H#koϑ^u \jV/Me\4=.> кhdRqE60z*`5.RA3ŸRqqx0xXw슇2lp(jzЫ#761gNj(6;ޏo/ٻRd#HBJhp|45dw{|u5!?ts_>NxA]v=U>+~J~/HzЅYɝrwo~_35}w/2ܻ`kq>Ɨk+חY`Ψ⋬T|!^Ӹh-Le.'#q;g{m-=<WEftm.emlC}AXxe'66|Baj,_ (j$"g]=]wuwey]pog J0vyeT_3ߛO+/U$h<(/x`Ҍ/3gKEoNԳ`7f`*(_/d狿VXhQ]$`׼0&c*s [Y`_ GC+F==3B݋4/y(|nbףE+sjHt.ꔋOBHrobDєED6JuzFGi*SF|W$\.%6!`BOYjViǬQFYJy/k-]=}6gau%oϜ0DklyVHuDce`xn'ËmEc[lyE60'@K&0E"(-*- @%m ]BZ/މkX҈UF: lj'+1zGe5I۩{U)ץ̎=_[$G`ѥD򊜄QVpeq_C ƅW{3~v4K>\AO+ =Q`P&09R20?0Es:'RP0Eg8Fx =O@0] xpTa9] ue889N;)>h:?6Ɋ .~ ٹ)d`eoJn;:;ߝk F '4[<걯NKw(&1ȊtipSu}yxt:p0-AQ6۱g{iTuQ۬8ÃSSL}P$!1Y FF0Ї_(XIEYwt4oJPI6Am"ssxh#}~\z<+A_Eu;sS؎;><~w>|~1&_LJ߀فӀ(|s|\[\ՠb7VR9ɂ WS^%eSAuSZG7t^n/)F>`wҕu3_2?bCa&CXbܙ/Ya5%/.ů/e_k#]$}憒!X XN*0cc3уVY\h QdV⭡%>5H0N92ecNFl~&j)%RmEvi%V[+AmR5?>6*7&].L:%_.ѫ)OrI9)`nvR\a_QkFBIQN)93Z*j9˜ [V>(ti#MJ΍iIzym1T)b+t) ^ wE%KPUFK:r1MZ|OC6Kuqϗ<ކO2gNC$ DT@:Yvװ%. W__/E[Ѡ$mw7qzkLշS֪֠\-\[M4-&4A->hw?cchΨL[道叓Gzj?mqyyGToMdRxŽr6F/2x2»k7oڡa9$$V80Jmc}J!j7ܲ{5tZȯq!jRf{^!)'IU61 RJfI@ -25rfl"d3zJN O<|螇$e;NS"Fr8#S닿̆(IG^j*;a/ E)f?͍rgaW粝$LvvZ\Vx,'߿Mpiz̜Ǧ՟΅9+𪐮 he_BD60v$BrDze. K|ʤGER{ (Ìy.|Jײ~P4LcΗl%|딣\V[@j~@E,E~29q/+i?3qbarÿ:˗NPgMB룬T.|51mqDF Ӵ:f.1pƣOKs&.r؝ V/G\S;6TR[m?Kdv(R]M&?|v>}L|l=u6<]{6pP"^%aXp3:1D_~-xtL4/śvOAT>Tر&XqcGAq4R`eVԴGjMyo= q-lò흷*)= s^Y( 1aNMc NIDh{',QdGdTaK0XEL DјI&rXLWDIcJ"l%#I4$H.b)9&M ,+DJJfKckDcPeޮԿ7=|2ধ-ѶG#ƄT ?XR:w/\2XQl#I6J!۠LnL )&DV PkqUsmc KjfqpGca(XwT5IZ23ʘ$"VܾPm͐F}GkKrU cQ8n'ҎŨ"3BfLrɴB0ѹJ 0Es1*7rh~Vk>8^qi_wB B4c1-3m#'Ɯ A+5wވVukѬT{xοtSt~w7GC˯łZMi?G^jY lH00 K$=F1mkI1#FnXKp3"اS[LAp $J7h -9 `h6Lĭs3O«s925C78΋yh~Ѥ7olf얭4+EL!A]\lLGؘ. ۹kێ.ϲ ]rgڳ)>H,ցxvuJa+3bv@'( )02 8KQKM|NDd+MJ-7~xū"$clgBX.xSg–u KQ~r$Įq-v*6]%l7pLX-zJ`( vVHNUrDݲî$rRUmM`WkLX2eWϐ])$J%)vЕ]%l~KsdWPQ  o Jq'lkW K]}T\~Y'A&Es3m 7K_~nSzqZU>?zc8'\Es! r(I/ vQX}0ݤX.d/^낉e6ko3\`H+/cpa&Rs\  IE$Ma,- WrZu!U ,yËObFw<3{5 {[/>WlG_Q,*G<Hܛ=Q9oF)Nh(MsʈVhmBތBvZC̼k JbAŔ뒕fԤHh q.l&D$A6h~SjkZu<5ު(8+ i6 X)`-RV): D#+kuv?3^\ r\h)X8Q9lY0*A(wlKH*4މk8f4pG죑N=n*QESowTvlR{lYߦ^hLyf{$G`ѥ+rFY[V>89nz0i%M}^B'=2(bGقASDn8RP[ gr N[lBiatH]x/&̵:g٤ ֡H4N4AV-Ԕ%y]z|xw7ǿ}ͻcL߿+0 .\n@Apzo> 9>~wA.'hW ZI' ~3\MySn{|@MNrr$ Jo ;,SK13`wҕu3_2?3u3<7b+vg Ք%؀z:4P4+@Swm͍!UcNj.S~J9e{+S)\w(&9̤H7IG9$s5C7PB.: (Yo} hS8,|sc;A4LHLr)xUօ\yBe|,&(c(ZKGSJl;$v84|[8OlhlNm6CϻmA~P{Gv |Jzm'eWu\Ֆ,SP?d8LƯ2nFO*L̀+ͫK&M*p 7̇\X-fKP+I~Ss:z GF)ܣiܣDЬnCK Wa^ޠs{΄VSY6n.?@4Ui˴}=l_A%oCw #G7j|[ί6{n,k/-z!0/4lQ!DfN1PF1g2Q{Bqyqύ?XǗM\dAP e5R,$)Rb֎b`rBRċO*+Il`>) ^2s\&ry o5NTv5% i]D㸴 zcXOڤ x\POЙ/uݹE ϟh. :n*o_fkͿI'Z锶~*o<{În5Ίw߽ېλ{ya~Y$^7f&gN+?\CzK/nt~/_]ܽ_Y~::'s8"),U?8cRӈdQgYpިKj vNqxcgoOӿ HPIY`D?]cLm3G]kP4~*cyݛą:S@P}bڦDPUbwm0!\L9`M+ wCғ+ 9~^>5+v"$dՅW OP$ńd /Ͽw 4/3^RXd:kt6J)Ko(#ٯvyC1؏ç`4G6Ii]3A'bRhP8HcFb1FozC13µڴls3A|v1RTBk!1JO']4E,5qPO&lIȃ.&JH et֮[9'n_,@nVׯjwN*NLSYzئHA1 m מ+ P(U Á>El6-,,eR`,r=P&CR F(ЊBR*Kl؈S/Q|7j+RvʠBVs#.Ưt}%,Ӱ]_(vU{:BWѥ ttj{]ƶ].[m7p"rm .}#6ͦ7ŢQ(^vU!dyHEAG-:e=PXJXǣC+1ےCHP@ZEhFPyŝm-`-.F >u :! hu\R$G $SC9C2}5ۈ΃ͷ`W3;}گ9~7 s P` mc^h;FX.x6ew8gB &[ yäuG,(/o4GmTӲ+_]i3-"Z Z(\qJ_WP}l A+kEyG_G\6s ^af0/VjaNˑ<<#< wRVk(IXd( *T2%zr#J"Dtө9P1(d9{["J ^OFiFs$Grdeg ~n?LOd{ xvle}~evYUÅыUhуd*[:h%68=͆ԏnԠRo7DьFOf]9ͩ@h$$ >,|C9Ba4쑷tk>fwdWgW7'۸=Ssq]ؑIsRwsHO#{FnmlB3zqw&4-Z0MF^e |ZY((%d /!Gщ6iBڎ3mǙgԺh13ZXJ[H2Jr<"1pUfoߟ9JehȊA6AE=A,-c-S'֫o%ЀaR8nS/PE!#t)BCQA푥Ϊ[Lg3 ݅!vI'( Qw[g;Ԭo*;bo A=zSK~;\LM ̎&}Lu$& ü͆UnK$7˂`3d%䒔)MHP..LHitkYEQP5Kr%HO% ](b:$zM*ȷ䘸3ZrF).sD ٙ_~p[r=ܑՏ\33Ž<2k:I$y ̮^&CjT;<6`l.L$`Z?h Fb c[3!+JUkNet / @V !V[fbpԽ+q{6SflQ55륹9]*)8_"rRzN$AHg=6_^IJlԂD70A. 4i"+%=u2D 7=قSoYo+YjWB o)fqEYPdTm\jU]a^K٪-s6[46 Vʈ1x i{? n8 .YP  #*ڡYlͮv:. P@IF{(`zrಉE[Ȩp9k+)Pl S5R<4[odƇ{Y]rO-_¸O-Ӵ˲6"~Xz\t5Dp)vA qY9HKYn~`SsO:ê=L U݃g$s3QZXB/}H1j'm,VE˜^'մT@8f#<] _}qYŸ9u[.ﯺjqOd,0Z*r;Dm)VEB$xdkph=³x:Ӹu cx`SvDW0!:sY2` JZG/A S"$׶jsN?zsH^}7cB*n1;hJ5?zG(R)0\Pv^8R.Z=9`")eUxq<*gi=[<ޛd/ `Y#$JL)K1ـ5!Zӈp=^GZtq\xV\&dP*e-iZ%DQNE y7fw0uhfM ) Ptd׍\?^ɆRN|}wИw`L&VnIIYI+ @˶n)7#}X!7~|!meN2CkmԷe!?r؝{7_y L7$^nBF\;3=a,*wFx^鳶t%w:ob%}DϏzeC6>JϗtGx[Y.?V Wwbi2TBÛ  xHqh;>SFNmC!v9^h|QB0|"G(SF}2֩P=\e56TO4hf_66G]ENzw rw S dvz~M{|qNG|%HgMFl}w'g URnW{OTYyE7訖pG\ 18Kx6Mj _l;J[h"o31sȭt///s \hdbo^Pcp[*P*V:Q"u;BҞȹcݍNgv51}|L@dВq@%& s!ǚh Y488 *-6WcI٤pEΊA4%]w şY7ye5R*Ai/=l\?#_C_i0uUL2_MJ#)iˣYc`o^"EOfPa! ?]Q}y&fԩ\՛ t(k7$^zMl%zm1QOfJ7_5ȵ~|W4yZC"6.ZKK25;o(QfVڰlӢp~y`F-iL X)8R$X&|"3a" U D.T@l饷2_0B+/JAp#Po \errpz WV LWR0B62L;\e*3+#nv fc*U֬v3Yp|J l*lU&nv2F2\m•M:j1pBj)WJ=zpeVn\!l +Z79ǰ-\m:\;.\ӊਮ[hM"CS72[ÚYu|pQte V&8ԡIiɤH..Qh!޻Zyt ^sPWX|b"˂.UF:Щ sa#~lh>p^f|ڍ\cAht6xqD8@2q{sv~ y"-LiySԎ8ڙ0.l+.q̞ߑA,~hT: B&WzE)@UQ)]R+!{z<$42wo0)\m`@G֒ݕ'36"07h]Js1OZE߱R$%J.Xl5- Q>piԚ:\hMpolg<+L-MђvZȹc@צb _W3%P֊3x/>=.,S@^$jeYhэq5yT@ Op>#sgD ՂHߚvLuoM޻+2yO%2WVtQќuNIH+)WUp{j@xҁ-'ϦE .(ԕԁhד Lgsn8.3'A>T}$9g8/N] &Y™=5iFWhHFv7mlSwzXC}{q=[təd.vLYpݏEF1?i.AU&MT2U֪U'jJs:>ylBmEM~9#Jvgn(Y꣊E :Ndp)VJERYmK8JP'S+ksּ0\5iS uJ=a6gUOF2KVkKjJo/vZة`4>]jDmJYQvHYgp5O>U^Ҡskmb{&&Ԇ'->oeO q?jyQz{oKдuf{]RmB\aVH_RxWj޸6P7@ \a`)xk"&DHFE+=h9o/L"uFK~piSu_CT~y41qepq2;|ͯTwiȎhi:z&_)X^5D$.HGiEU^Ө-.)Eb57ֿd+w kjRL>$mYJ%ٖMٲN̨t)^<*g1wSJRa,,γI VHlIJ {͑ؤc`uD)1S>J. e5喸GЄW4oS)hlR>YcNKm &y@(sٛ5ʶDF)(-ٕҨc'c3^8ѣ[׎3$7T́14:1j%!ݘ,D̃skD!&@@`:%"=)w{쐜 YWGHs& Sdph K HI#~W+p Wg?zhN>.g§PA.>4vhкxt-lEv-ڛ4t3n#:z%t z莎]!j<%xc%uH)^U␢jq~>Jr$Hz|GLmPJrea5(E@ Ec"UgZIBt$;0dX(LWH@-^bTL -k,g)U;cn蕛ҫO :>vE?܀0_7yT!|~(8pjpAStQYH#lAu]MhL߰Pϥ@ઊ.ZGƅֽջdDGhvBSMMS68u~27XÔ-m 8~<\Z,b&bU⿭>gYx]_ uh{/o.Ձ5PyȚ4xݲ[߄c߼c߼cIW#;m 8RVT$M;G*r* :˯Z#٬o _4?ˑ2F#%btEփfQUjJA1G)EyW=u~}_C"to^߀r1?s꺭&\x6d>u`5%9`;ȟ gEGV3+nR(Zi0 ]A#{/}T*ƒ&rYym~icKEUS-M"Vh\6*kQ5DA5AQ7qV+m7>ee^J;#i.ӛ S*/y+Ag-I1aV< av=51eΐew.ʆQG3fH9#sGERPՎ<YEi4H3Ķ/5(3YS.ǬcJT@lK'H>m s@.m[kevﷲ4nQ`u3W«^io1C›iV34i7fhR86f3hoߘ"OWÃK5QN ̃aE,C':$o-{d#>Q֝(Xkl8IC+TLK. } `Цx罟yFJ;s5;^7['tw.CsɃ~|__.L(McN+0A0F5h]Dm0DR+Q|H[?1s2k?@44D[L+ t#{O=JEZXV5#`BŒJ\X[b.{b{vT͙b#|).z|u Q~e?/&ǁPՃqX d)2%:@(2 h^0x\yʥrL+L-S  sū̹N@}Lh6[͖"%_@{_:ئqjk(oŵ3|_Ϗ, Tp]ѹpetgcw7 `i&jEΧC9x@Qm=ɘrfl0a\&8_L$}n.xX#MނTo/Կǵ)Q٘5W% sQd_'BFFIJpet EugϧwkyJcEJ)*>,Wz9X% rRImMths83?}mmsutJ/Ý=,sG\x}6;:>~OSL O+R#?%7$H0@pIKcq͉X dlAgGTMT-BIؔGJrWkWdwՠB\Q`CVKMQO ;#nxniEEs3:[ӬU/?󛷶Uph/ ʅgYW잩ovყ^̹o|{)2CI7\JZYĐsmowPF#=>͟b {O3 ymЪnj/uL* ½kw藴giXl:_M^^8:,jѮ17=az9O&0>̟e;2<8@zv5=Ӗt&a3UeBއh}idZ^i鐧rQ˵ī,SjEuWnEza¼ `TNZF>ex%d}Cai+~lv6'n ⳻Y3&eaЉ&oxzv-MEⷳsۇW|73<6io|i9}K1|\hZM6Vc j |-_Ԟk貴![^ R R_NNۻjX@|="EE\kdrb9C SLua]tJ!%=*/ E͖bV?k-Qill-#z&ں^pq*k~-bý맂VmÏPWPs؛fz7qAMZoNy2/ܠ9/HE4x|=YK ##;hN/eyʑ@ œ %S#%:Ԡ,`䢲YkgWV ; u Ĥa띟dR^3#bD!ѾF*;LRdK6Xn5zHɨhM諨 *Dؓ5fv|v}a $Ju'W(س8eh=L2=bSN&G_㳍TߎV"eYRGoK^Bmi(99jdL ul=*HG,j#3ftH$lU!.^{]Abz[n췌J7[me<Bg[-fi@:txؚ8wac`<DlM?ED2";Dܺ%EK9@E%Ǥ8sU.L &_Lyvv * h' ٬4Qg06%l#ZPj8{h~X\^ɏess̽2yڀ0wFF|v*pF`\)t!!Yb9'6:`4P-4*XVhPNOʄ)]s5qvhV9r3e(>gEsJk ^,C Ɖ;oOj R$e88W$}@oʤPPX?*2M^wu]ᆱprgTK5^?+Wu%*?%he1 djTiEq *6=->=w#Y'W_X3 ۓum)%2I*r7oꚇN3_A 6|#,ENoU{Y"ŋq?AdSv"30|B.OZd% kdҺAH%`EḞTလsj~Ÿ{S.9EEU'_(_h5/Bկ R[üAhi5Ͻ~2;n-E^8aGQəT\NuROɝ%{y I%%29dpgBxj ه")S3I_/CT8 *br&t}};kV_9NdI> Kv}"aSW楄8'B¡__% ?ߪw6p}>~4ZSԽ*ǭG.ʣ4jܒ*g9(7zS>p'C|7Xg)g5u&W&&,&ýPwg w UK2X.!䂍,$N+N[]kKJ})S)(B*4\j\Js 3x!dR1!iEc9ZU֚8;fI=KyR87>RR%l4>hTDbɖ`2d*)p+0,c⎑KB7`鱛 s5Lv1c2^>EReStYFtXR= e1ǘє9ц2hOZ*0 hdg%Jtdg%׍cc4LS6bAvcʖMlpٲ1ҥpٖE3=Po;e7/i< DʳInagLh'@&C$>G F&-R']^x; ־>|Ye-x&A($\œ88LB2ʴslwN䏂l M,'K% ?omTDB6*ZFy>Ƕ-qv97C:^ڠOL# 6P\^ޞqvVSgZr-89vJ?{}3G O'׶&j~>v릨IyhQQط1Cms8g<  6j~+Q22Ib ,UP<3$;_=eiʚ*rMSvѝ1j[rn}uJ*ymeNj>Zgc}FS8. t=V'9D=݉{62H,3: ,"(}pvX,Tt4iq!ЂKH+8ϴgR g nMu6+͋e\F,R;+d~u}CކUYenڞgqmDtګbOߊ*ٚ'*C 9A(VPOVn=}H!g9)l%#)Y+6ԣƤ 8> 9#`13ÒC9;rf6i%^̼3[4M=N>X(W}7x[3$Z3ZEn_Pv37hM1@~lZ:VnJG=sB3Ulgtޕ]#N, Gk\14ؤ!(.r~Y[2eḴf#!%BHu)ҌTXV2pg{Dn&& ʍMpRfJeuvfmSɡr$Qg>*Ӿ[/6hNN`$Kx'is,:9? #~NGDQ}A, 54& UMPtRXn\JJfn+)EL$3Y,\[gG)6j4m"?a#7pMDn, ymXZC OW$"읁MΦ;;W=M:m|{; Uz>&\!B۫.~2 B.o3EK7/ɴu;*]N=m6bI%L[wvUu+PPR07=Cuw x:`;*.7-9oI6l bYmmT09v^== 1!fcKin^ߔ{i@w9_s4d>|# .bfoH3t~.?A5reDxRUƗ~<}`x9ؐPȾ6q쪞1͛e_a_d8Ԙy{/k:P%5;67SiGRY(BG=9jL/i:1ҒI+"ˮt$@x-R,}u2zjb/jI\R5)I-R*Tβ$PN293 YD2%F;@Q^iyNzH8g}RM7Yfm0wT﹨Q~0<`E\y4Vz~0R2qFڷLWOW]ɀ`{U[SI0EWbK)+9MQܲG^kz/or>//gmv1ǰU#g>)_}>)fj{/μ~sE|܅8UCgo\#ju8B"F_%Ëm%3ÛDJ5({S3==5U_WUWWɝ68c"6NhPP-\ZC2-!. F"ƠU wVTY(Z¡'BI>LG5)h%!}%5rnzW lITr~WgBtaMQslWtY\ ~68 mSOmj`ZztRJG+4KUg=vcg=HTIK &4ZF|0^'ꀖQTp…(YT B%cVn8;.sV?q~~zm|t*h*T1z?}F, 5LXbb%a9fp)]oTٔД{ R ٓ`[3H4g"!OHjQm'=V d?o`5ۚL(~vӯgS׹^gnޤ[jyU G V-DU8'RE !81a[T`E+!։6!@F*oy@FR7uδQQ!y^S9g`=3Ǜ)8FM>mcqxp 6c=Sfr"nh|x4LD r]ĬĹ gl84Ƽvg8z6BO<_X\O)j?pZ&ы (THByAސ3UJ#ǣzcS. C ٜ{<h8"K=:IY֤~X~dv[ !+Cu'mˮo'p<{"[җ,M]ȎQZ))yǃd(l!xᴷqAOTM`0)MNPk&& Q/PN*'f"S!14w)sZV3˗ waPeGzzCgnqE*xj2G31X dGHFJ 3{.GPKoo]mӻzgļ+'!n}Sl,ϰ WTnӜꟽ|qQM3P=*C1ԇ0V}rJq&Ë<{[ /,#4Jh9w?fTB8N?qv*~=94X;tO^]LՏxUѹ_wIHQnfn8HB$]"\gwSі T1"QuO#&Rc-hcˉ]^|!VJVwp@8N 5Q5Ǹ"GRT^'JQrn)] '+rl~&BrA6sS94.2jY+t'.ƢfqK\l4"NqћWI8ugRԿx~>CoUlj%Kߛʕr383Mofw=x-OwRquG31I&ig:h©[wmvAn\ISn6{er⬳.;z|ę6W3r31No^xIE^s)P-­`M!()Tb B S%,|}-E\5 gq6!D=br<)As/$!KM}ΐځ^+90r ʓqkoxɖ׎_Tö픲Eָb(,ד.tI6(Ɇ1v!|?uׯLϸ˜6mɹ' RhJ+=/io]o0ۆ9nRكHޱ(zg,KPA82$ DJj\)@ 1P ༷T Uhuar)U2% IKM2C>$dJL+ رl ߒ_,£h%7FPeZȹ]P *Gվ|OqPdW+lXr0]V롤[wX''N 6vy[SeAgzcɊ**oqDԨăbbLsxM NJ,OS]Ŕ =1"`b!@etLֳu$+T9#c{JkXg슅e,=f1fxY&.6w;~\ElD$$v( *j^4qo QmK8-|'4dcEF !MtP+5hcSN&JH6 ڴO-rnGl?%PP5x,j#lN( %8z}f˕'uМrc\+nm$vIZȨRVtk"ڒ(#b^p>&I@kWk܎Q?x(슈eDT"vxKu: J.qg̫](T*S^,9JHH!J A NmRKg<#[*,hI3T2"F p=x?~PζdW\-pśqR  4KjKxޏJFB2&WDʝD})u\<m\ζ E?>L#'ĚE9MSpلv/7)<.Qa&tWrIec7~y> Ж?xr<6֧b8klpA0lX|qFz!RҖ:ׇ1JЇŞ; '!W64)‡NXJBQNFtzqbB5. RAm*pc9+[/zD&9|*R)ޖɼoļs?wR%0-?>ŲȎe`ٻ#͜cmԒEXZ-EaNe{M^n2+9Y6sK*̹x\ EZ.rTFxϲlj0PCJ4]iTf?ts|úpdJzĥGIſ!cesPͯXڮMAmrF Ӳң;r8FSʼn\8pϝ1M(lT J de-2* |BWdWiLkѢԆit҆ BSYPgm<*C!j+m؟驡&HBig 1#X#LKBB5؁5-M_jiڪBT$F8Rz'B-;$)*F-@Gp2N|0qNr|N3We194wUrSp:Iu;>;=HC?WϚ#KU R9SU%\F05SU] f/?~5NZ;=ǕŸ.9L'&37l!T\&!׿|kuw׾fQB6V55U;]k!_]dl/CHR- ^mղIbH$kti[cY!gx><I9]MO[7-47m*i%e^Ov59Mlnevl^;?]~}uKBJ\MU"URٌ Vq7}[:uz[vܤԉo] BۑLӔܘwne 5J`QA<A1T`26qgVY\h QdVPz)kl{r`*rF+e$12dcNGF ~&j)%RmEɲWI[A9껕mwlWyyjݜ͙ Iv]дW{hs̼ἅ46>:ݒa~aÀ1T)b+t)ū ^ wE%K J*!&"10Xj~: sθnp ;_B3ri39t\ ܻ<7?MkNv+ЧRe4o5G@VL GnN-z?B\  %H0> & C: J?@\p¿C$-WIJZzp%0EŮ@`U \%i9wJR 3+*` e \炻H)j•"Hq@ppK:iW 09Ws^Cҿ0Tzљsa=ל;b ux%=J#AH  LYklRER[ZDso^|pro|;2`Ƹc|өb#HU15>u;VV+Efo/n./Vs|4q'`]+7ɢ>E9:e 9IGlKVC), -({9B{JXs+Yn緽̖GK XQ1"BF̭D9łG"cYhPNq,zN*h%{g-:aFhNPIssEސ l|= \-]^xyig]I`rk L>ObZujGDCG`*j+x2xIΜ?|)9@epp قe k8/:`)-milKckfb5OpK N UO,%qC(AhsH eb-K/>ؗ ֣&[dsF[hꉠVrEk=ck=(J)t)NYAXFcFeid垔sL(QmGsh[jEw9lVJHCF{fg3.؏p{e|f~^@?aX_o[,b>DCJe[8ݙlnԒI697y70C={ؚAy,  ,u>jc J%15w2py aؚۭ&}vz]{JMyxue,zz_YN!?sp[=,X* ;\EM#v!LC_#x͙BڰX Z`x``LR&Jb1hcu #}FZŀ3P"|"HrŔQ,Yydc3Ex))᠜nb΂wD-^&V#s6K`l P*I,\-& TsCjm|C5 .%M,B*RlY'- KxozZyjȣS4c`?^`)e -qb& 9"OItݸ^K8ZǏJ"g4P?#pˑ BY,j FW!D7FoӨr-TT *QUqZ~<aH*PJw;GWiQ{~wC!գͪ3^t0$r7 U(;I"G qP_ U:}ofU+J< _wM}2>Ds[]1% a7u/7ޫNLߧ0d?/ӫTq;M }9PD]ɥ ڶ}rx|pXI<+~<m 1NJh؋9"{ɱ$Q {">aWY5AyJ;GR 4~:r s`1F: lp6HFZ{DK6ûZ(6 M#bBrE3B qlHd3⨥O[* 3iJd^E I  h@.(|8} ,($S{+%ZvgE&fE @1-{mqlHb{tlwiDsG] !. 6xȘ2XrsuP&0p15yC %#l-J+B))&=)Q[Mn ,vLʍٌJ6,62n,3 =̛q|qyUƽ #$x&\1"D+4K5+١X3JE!s6xE#gcN`=G@9mV&\4B`RB"mpjT,eq(ض*QaAKVQaDliĩviVu6%jU-.^F#Q,QƲD# EPNKT(cB8I ԗRŭæcU< Y_a &G؉co*H6'q3ޏ6EٺssÎeC/G#0ΡD`*R:ޠ9u{|rG|&JCg}9\7fM>;?.bR &jf6m80(3Y`j][yp2ˡ~Xp ٶeNxiIb ]M+ 5`s*E `<$f8͍N6и28Mb$qvrZom mklȄa\M)gAײNX@yK1RmKG࢕DŽW0S!մ"}k!y{2Z6F!_)% &D`S L".)L~ƨ| +AM= y.4JQ#e [ଶq.DzQ;nh O0#瞶ӌϷ&l+Os 2%4 G~R{Iea]-O BnK9.W0EQyYZKGIPب"tQ٘lB% U;dUl|B9~M5v3%#SJAi4 TDD%YȻmqNJqOb/օ8n/*1ae?%QM.TVshZ6ˋM"R휉 `El#?viz>$A3Yg#$#ޣ"RҞ:.",F.<Ԭ x9 Wc:C^dWxE+j(_b&ĽD-I]pf П?f.[o(ʐh;Po0ŎТ|yW$窔B0"ߥ>c~( h`}0Ì??/nnLTO e0T佽A|yQq͠{`\g TQB9a$};3׽cה{$rWQ (9*^(O 3rHWI7S*f>fE/~v>Ƌb`+0*4:<< /ƠQ#?wG`sw甩`\խa׊ 0}~+yGӢ_@\ð]R<{N]q~8~7~:ooz2}_x:{8s:%$HM¯"`NݽVB747mkiT .voVVmJ8?*ޝ~|V?rD ׽POllegT+*v*c,PAFa5%`پ|WzV1a9$J$꣒.tPˉ6J) QƓh'NSm qb^Ymcer鵷FkxH *:L\4Dd@#s6Nq2%4r3.7+&Rf[j-ǘ*/bʹ@j}6 5ʇiR(8q"5cxka3?eb'wդ}vQʏ| ųQs"}md<\/o~R;uyZ\1EE$hJV.zDJΪE 93]^H/~Q/%MDtVHIѫ8˟#zgtS]V@ͱ† ]x__~Z凓;Tn5[p7(^[cCzsnF]֦/ fO:Y`Vz TDO k; $Rr] ؐIA3}S%2W;_~E(H-ښ׋w&b2 rƒ% wɣHCAVJ Qp.l[|RYgBb) SM`J-.Y%S&%Xbm!+Ff<(ۿwoV:'hVXNdv/c9OMbhJ!Ml]`={>>'bcSc4z,ӵ:{p=cmt*D&PQqJRTtΨ\G"lNZ/ƒoE{0BFfM.hTB1L',0#" Rhu>-XT'cN3e hq1U6M"Pe 8>%+U2|0>/8%ˢ4=Z6he1,#XOxy~ГOy>|tݼűՎ6 k(+:R JJ_k!ꈷh(5$]ZCrvόl=8GmcA9jEZ4:* L,\@7Q*%=8ƌK唘*FZy>X٨&bS6ɧ9ԸԻX-ͮctu/M mE"*p$HASI$Ռ5{,L$.}b"ڇL q&e(#hh9* ʒj=kX#n!i]m;^]tٕm=tyrZh~Ӣrf-5|m;W&`_UjkD̖fn΢[wbc=>{}hRKQsegTirB8,(T PKz]#zbiUK 4sM8_*}ˇϞZ_e2iT"0\. τKEID|(O]+D{dw&#Og=k0p{n~7oUTq[W&f# e6rV*~>փPG*qT^aUKnX.}$bW"Sgg~,F;~XVTdFm2A7 #0ΡTWVJW mT@]:qQi(qXB'N0%9%tֱ$IOn6?j4؜ף7QJ^%w?EvKJR*3Tc.EDscc c#$gSGH,sN_ЄiM1rԯ7?Kv!gPq-y#ӝ;7sien"s.tC/;ytg2/#fyy'02\!(>خUl}xRP3+ !B* ;WZE2Zp J !(2+p|*S9-\=9!B;WHV+prۭ߻RBh#\ia;WHѻWZ漓WZɶ2p Hj䌘W\)vZö~=S)Zcہ+sϩ7mp0r͆Dz3p0*Ֆ ijթ-8|zg*R+vTΝp Sν+0jug?ˌZFX+s#?9lAULr !x(A7=_F?QPfFnJX[_*kY+pɃ0I(O^,/1}m Y0 wIIp%X#KcC*spM5ڙ].C$ΕEVq#YaEe&=`6 ryA_medI+qO5IɒmJe3|DE믺BbeU09 'Yå%?%,#T@l奷%A0罥Behu%,wG{ 7g 6dًsxcWeId|k6J n!XOL<fd-C[<=ڂgɪ ecջ-\5\q\f,8|:'r%| 78>r]i}0DliO`lZQߦx8XYUo X[]F1N`4ɀ0,]cf_Cq{wzͻyv<57wym淍5f8ﺙIJHGe-[ 5ռ֦эl[ҞR=2J*v4Ju>ϣ9>D=88Uz- Y纕R+eBN%QD@NJx4$hbϳɳ:GGZvIOP22Zi|qT^Y3|OY˹-NjE?W ҳŘ sik>u0TUVGUe\L ޓťwq9AH">W!# !"3H$RllŸ([('..G޲.WgwdvEC;rKo>&&i:UA' 8U*oY~s5-@c3B*郎@S F}Oc{;jF+3}esGsyè,&1.((KhFhX&bQ)M@B-d85>K%wOW9$wM8CEri J"0`$'a}` jpT&o ,ZG ?,"f8QNG+ i kp1qvk<#쬝:oϽXMm~p][dY2yڽqa}mI>zLSjo^LލO_Z^d2j84 *²&TuO6?M)婑GZ/'7MF#H$E=#lJ"O@R!GHj6z9A4y70pockN_R?FCߑm08],mߤ֛>8φa^u{lz#,E3om *xGdDˍ]xѢ扆x/*0ϢDGA<&"@ 9s%*XbRCLO6z$,9/( LfCO'!*% xJDCikY0D/$ !gz)<0N<A3#NjzcS. C 921&y0+q6/EzuXc]ŴvaWeݤ:/(A2%Җpz!  mt <ѦHcVJJe` Y4: vD8@2q{6&izp$i3C sitq˯smyi^8 ~|;Z6 .}I?ptOh?&~/5?{e>㪯inc R3&WfƱMkn {[]؜mx5x+_o?}o=|1oq1}1~6YLd:?=a__/M޽\Ql2;v8w_NǷ~-曆wcV|꒒3F<ܬlZ"hވpyu= +ՃrJ]S+:cXQo8wnb\:?a!YAp1%9?ݰ&;i5`CdwI=/B5$]~yxStY6m"$w5z8OߵcKq7 /xג [Ԟ~nx]Y[XBXIW])jrLW'}2Luٙ9ʢCu )$£%%ނJLyaA(9pQOy(C6]WOx?-үPH E}qHL *:8P,QO469^zE@/Ϋ9,eǠ_K:si}3^<.ʪ “Ǫ}Pʡדg6XR>lS-b"F T˜WK}Y"DH=UTBSVYyE#xx !ڟ`9|@ٓޱ(zg,KPA82$ DJO(:0h8$ L } {KʐbE)R!(*LxgV>5{qG'ElD"$v( 5yE7AXjH'NK($Pl!BJ%Q3x16<=ªbM$ٍ~2k&橠vYzDl%"jp< 4 ,͖+O9Ri3˔/V(lA$vIZ @+:pЈ5mITOfBIFШJ)&nZ)tGUyXw*|z~NN{ҥ8/a)R#Rtʯr*AQIwKzl{_N1q ڔ]g(15s &eJ,L yMwCD+]Fh{ y$m~&rgA`k0`ApvQTƒSC#~M)(ٽ#p{wt]@x[8֛1_趹d>D:L€;q1E !ZFRdg4Ihs!HD6E,D\&2F"&763s V4Ѻ)ZKogOi5O&6oOk˘Y!XqviOĦIIﳥtN.tM`4UxʅAP;c'$'P٨DJA;H. CJ_xj3a3ǫ=U L+`ZKІit҆?{Ƒ8@ A8xصܮ!Z"L -/߯z/Q(QmE Uwupؼ2Xjp(ڜ"QwSy *hc VzDzj ҃P}B ycL&n-$$At^i/tRh lK Qw HR@U.~-@GpFpN2qNrbī2ј\XS%'95Ke^xߒ2/PF us甇;1*l97k'R2! H6\`:#7tQ>[eNj9a:o WT^uQ‘=~FZtj$bw'\ln`nL3\xV#sҭR{|Ҍ+q]X(B6^' lCAȺXے[:] [Lke3r_~S![~;hf;sVm[vr]kjB_Gy>yE|vhov tXŮ\~ByYjn;8Av{wG>gGzٯ(3Gghup0t֐$5 ooE9nZKV޴in榍mz2۵\5^4evkfv ?Ͼ<(khz'o:-&z b?As&;*KEp&TYNcB @@KEaI^[m;t?d9WG\9 !uRDTJERYmK8JQ+ky~7F &8k<$uJ=a6_OF2KVkSjJ6vZyT[㉭6޷U0Զ;8lWymRs4 \az& i?tRbW80?q]̳-k#{%MR8Jb*sZ w'2Vl=#/0Wl K[8I;Ǎ4ᣱH99hSFEL1Vz(9o#ù^2faW˓Pz(zlbbe-Btqk-]6릾߫Usd"|5l]a^oY|Xkm Zah/ċ_6;@HKDw \o)~RX0\hKAZ@qǏ>lQ-i!uL,2mP)Y.  уWi Q􅜊 Mlt@4̫Ζrrm4DJR0HuLy:i6Adi!&8rd;@4DYƃ2i" Z OڄQY㷍\A~)&FE,(Q(P}>Ɂ8*j_?FEr`z?A8.qgyUΉ/F=AA ]_~)޸һ~SZ&ʸԺKQ?DXx}Lg…?ϺSg_Y= 9c6 ӈ#dKJA—Ӗ(bTb`5U1`qBcd96ye /Ce(p^27B xp˅Hk[Bح^>,'!g-@QŻW޼/Fctq5݋hiSKMi侩?;,;(~? /nm?;=E@uu."l#ptZQ4xou."-.)Eb@"T m]ҏ)΀L (8G 1aE"`2$,`V R~렝@p9@H'Bs-s|KAtIFrkڝ z2f4NhC68%@K%6YhB^Fq*="ByHϦ~߲vz 5=Z1{1&ǖ$i:/:؆Rqj#I; B:cJ!,=Wz%04lR Vx.p־6 se}ϓJ[)+ >&We9) cD-}W@L3?NңO(z۬2`m9:yV'O3I0~>Mm bZCGA|::'."rsZب"Ak%tz\(HTF$S:KrEjGSKX}옔5*2M 9&rADJh^8˘IcN(j;Y&o6ͻT-VX_>+зI#&: )$x+-BXj UD}$wvАH1YCs&Q8S1t<Me kY[#gPdm o/Y:0ee[+ mX8>~ (Gߴ "\Yjz~w}UԝV/Lʫ?}+~|i2Kϙ+Qݜq~ ٽ-O(,HFpmTsRz!pBSt./ ׵P'6WwkjgyçFWy\;2j4WW*Qrzr4=|(ONx[ zdw'i>fG͹ ( |oȿJRvK8sϨt=Wɓ|_zYIU ^؝W Wzp%Ѱrj M@wŐ`A)!(A\dS`Qe@\P1Bgl5UN#WE2જЬN)@3MyK@UK=RzMVn5G Ů|,3>}m&Mօ Pə>ЛRm&LiEs^`rq+&wMku_ Xr4biR 0k[h⭓Em^yI\Pi4J lD- jh8Ήm[VYHE$*mS΢nD &$%ᶦWR+!:lZ ~8r(39gp2,(nD)RYLpk)#Y2RJRW:IB)"$I{DRȹB(Z*#5]Q#Cs!G ՆL9Ǯpp8rpPty ZDTIbJF`vB+쳒3b e09*GyoA†QϮ\N)Tڧnƃ pB( 2NAQzrZz'^ؠ\UB;P^_o%f 蛂OB肃) y+Q5$)J.Y#4-oA¦QGٗZӦ> ʄK)p 8q4h}cJ>w8!)4)VQkѵL[&md,\I )o.7D 8D,F9Auk\Dk%]pZYJhZG HN:ZiBmyIqǻhtNk:b Ԕm+8)2DQQ@l@kEkB\%)LXul` 'd3ll0wxz E*"3IIil12`[f׭sTi<ت6:h!Ƞ/C%Op j&}[:h4 C E+}cl$TWfeY\ A$b`i4dÜӨ6@"KN {P2 dJBDJsC ` өY2Y`"V(+R `&L(uo;C ;0;@:9,g !./5ȳBnTzCJC+d2aޛoCDaȇv*;"A aK-܋)PI0t؀*VQr 3Ip L-VJ (Uv e*3jƀ(.HNu0؋)X5*f}E 0ABYrA Dbkj{PP}^R64LbqѳڪEBb|UTXɑx 5 %C  e06ueH WE7 U+cWuDMdPg:c ֛BQK3fHNV!*)_,K8G!^Dv %):J{a2YWҥ`rD@KV00U8&q 'jK *,J0F<(L#/$b{Wȼ"P>X\Lt~BGƁ:SPHm( AJt+Sx$ 3`G[}¢"Yԏ/XA}( Ӭ(a`dNI.+IHl' +(M>sWpP)Q^"_(r1#J冇ls"AY.2!S :G T", wB)7Ѓ#F7l֫Ű$>Gw dE^H"\x>7\N.;?$ VV0Zة !jYPcx)U,JňOg=kU6Vzf0?yڰ(RVУgɈSZ]FύLGj F5h*D x݈/byP`ҧ+aAp-c|ɑ燝rMnNw{Mg TrGQ}0kJ!#_X@a33¡FXV2С M@XD}/0""5>d֞Clָ[XfP 9$bR竈p0زK1!;Fff%Ф1%܄$"Ȱ?tĒC' \NklB9KLרvEBPީED8ØՀ]- 5۵q>_nQ*C%21cemldX_9 T`풳7(N/_>X(Ŕz9E!$"b]c:֮ͮ&&O )&&@‚@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DLϖ *ksDL ̵:&PcaA#˞(J)81>C&@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DLϗ $$<嘘@B*掇 T\G3 S@ 5%&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL beG^H٣aAZ<RjCLϑ dIb@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1>&лK5nhi/G'/fuiul/_-Jɋn(of˻$~ʙ p]|ꚠ ˵鷯KYίO_.k=\b>46U1!Uc)r368/f_n_G"ng;?\W <}3 Wp7jшkXcz6ܗ7L?uݨc_2\6 h-H*r_Q?g7>k3wd槣[띔=,-_]"s;oF_ku$}W`0_[bn[z?cϖ6;VF^6~79:#ff^n o}7JӺ|'ݣGu 2D͛QW,û.>$InvFmuIn^puTR:qTs:q#*j&ōT0P/,2яJ5 n0?xէ6^F:Yk j!-JBꩼZ+\oЍ˗o7ǿ!?Y[ֱ烺M_>m {>cn뭓C-77B(ZriN7·V;}V/vn>?wP }ZjǓWm;\$,nK6,tu7O?0XY & _VKKVEOXJQ qZUHvu5&d'Ehw|Em.PЅ?pS&v˛fqv_3' +IUeRY`~Njj%W7l,DAIq949 xOw^B_g/Uӧ{!mLl%ȿMkgy:Ej5NΉT6vki1Ne*ӺXWfKa}|OiKo (v&ʚ汒Otwykveq4;8wJܖ 70 Xhf9_oJ Q,B]VDm4'nv̳Աǜ |w:cR>HSmc3ov]qG'WWmpjMj:W}W^ }!#7>@P#kvoͼVrEό# %9[KLhL?L(hcf#S4H uro`:f`C%M6VdiI=yy'a vNb<0tah^=oUml-\>ynY5 EׇO.w3<9L'(yxhƛ7b1m|5ae+磿ݵ6zYOf˭pjl͵d_ܑ[~6Ϲ >$4Fjwhp[# 4yr5mlÏ5\Mg{ /W犳HT28g/MPp6n`ņ҄I/@|S{> ֠`e}aӧS;kLGK:elppyE=`as/ `p(18R҆e'X: PJ EnRMR=;TEH9Ұb,Az!3tL>;U04> &kiɆcod)'܉"n?l&}n=E746.g\nn ?|o&Mz.X_v hVŐ~7]T<0яɻ.+}x]I^ĴzHl\3c.L^Ӌz%4,2Nߜ.eǫka%(XW|!1V̛mb֕B|#n~Mvf?UK*nJXQz#g?7(adZz^wl(jgݧ{v˟J.lzju]~AxnR(rVVwtkw? _,_>fFIukdL'qv\&`foj ]|*6;-cZ2joKJnr_3b3ݎB;6gp\f6FͻMj&Ϧomsr9Y׹nQVwmKprZ{4m$^mղx;KBɢ,۴-;T&%3γ3h]gtgUEpg0\?+B* (}NrSsޠ:~ݏӿ^_~w:|w?o`RekCx{p?y[Kں5ovkշVUbz¯GneS6oܮD糏/^pӫBS҅ŬU3%0l*&+27Z,;N*7MeH&cl? w:J_iVSIf,c aCw&z@EυڡEVan%^q鼼ϻ6 o~ǚc1x1qVuy6\3Epu-":zFL~&j)%|RmE )lt0ZK h˥J7Gl+x෷mnީ$s@9__FK*m@1T)b+t Gjd~9HysQ 6BFM0rʘӮV?P#T6da>8l풁C @ljxɥkW }ӧbЀh׭+lv[Wc ӛ,BM NZy5=UV*k fO(.ɬ_U% 蝇zEOep7oUkwkWk m6փ -pj[e/"ZÃ-Ct c4Wɜz+=h5z 8m8w##E𦥏XaΉ&2)^9pe1#3RBFxrCJbgKԖiB2 YD#hؿE81hvgY1'7Y+&ǟ?>7ì0p={e쫯Z{V{>mF} fѰsjo3ϫjl\:ڞlx's /47e#`2g+}s%oJD6Q%B-8"BCTٹmy~7 EE;wN$\엽,׿6-N0ZHY* A~4gg̸zN)曛CD_Cbv_Ìqv߻? c?d% HX21mGF묂:1Mf:&!K>qrf' "ˉUh/]Tz5e9`Hk_ZݪfYR).S(|ɷ{oN|[uVMY{V#f&^7ǒ[9Hry>؋9S!ao ?;NKEk+" ih"`Vb/5& 8El{.[]/}w\bW3JC FX࠭%`aHĈ a15ig \%( p*0XEL LјI&r޹QXD)cJ"l%#I4$H.b)9&M `,#DJJq;7iݓ;h ښsFDk4SrlI..h "Ic1F1sɭ`)$Ť#"DDQo [ la("YXg G& ,0c $*pg jU! |O[§P&v*G7: Fz]ߞ[45nѕ3c di&烼udfrɴB0B9rh"bQFn-:6wu!9 r >vB B4y@ IՖMcN ;oDؑ0b EcD2Y+(+5eW1(b FрG!eLDep\]XyRt`šaZ_j5*[>OD[o6`aHzVs(c >"֒zbpGN"ݰ0,Rg2DO"p$J7d-9 ঒h6,ĭ j!o.I:j2Q芅NDbc.G-_5:>Knt-le=?m[[9fsY04\$h;Q,^',A('8(*5bK,xns˸bIeƸE,p]+txx[@[\3c=W O3?T6H <jؚ 4p턷ᝢ w0}ji_or -إ|7߬cZj(S.I+Z b}{8O_$ _>L0Y[p 9-rRln%9Y;c{=,KzE؛CJS %Cj1Qjc bye:99&r6hbR1 Ƹ#aR0XbKc۸\+W`{5> +fZ@+}?(}e5_z#!LJYx|ěo ݁¸t+vY6,*,? TS b38Un"Ւ[e,:u~OQ 8y*3tn8+0Փk~yb=:*bD:(gE 9HNXA[.B0&v:Au(8`1wat ʽ3-:aFhNPƖ g^~y{jϳRбvz3Cw*X}G=?~n=lZ0>^UWI[ū2g|ބ b5F\I@6gev3y$iv0Qki.j3J7EKn+:a`'nVvΗ4ڐ7ӦgVt'LEb|t &CRW}mFaEqN9;#HL`]Z$r>wpGqJB]8k9XlFI-`L͝ \u^GaX.4s(734[UYAl/n=fjg g FP5UwfF36XCgX5g i:p'!4C@10&Bp^qaU >m{DH`J^@.hqKNЁ5\Ӑr4UK$(uװ,hiц|zVՌJp>,8&̻*$MKH H_[,;@/Y)vp<9iT=yjGk`<ƀ~BHSyGAr[''MH9iVySAkG/J"g4PM>*~){SߓO)LN7z&騟7KKU|oN{O?$1{{ŋ?>Lt㼯fZ+Tud 6-|;<ϱ|lT;4&Oߖ 3ٳ^myGO/ zW8Ǡs`.?69׿8|g=|. SpawŋS3zg=Q>h7_`1=r:>Oߦy nR)Ƈ+3.:o.1'X}LU Ǔ!kC{K3ASPyrRnA>aH3kB) _`;ǟiQ!^}zμ URW $nXvSvt ٻ޶$W|]~T  LEzAPrIA+{%ie61H"]u~T/&:i+D&dzo7F-If3Ijuo78JpyM./Orڱե+H7OOƿxV(9Oȳ+~<;i_|ٺn׍''5UáD/&nO^I,imlImIH3cR -#MzVbz҆]W\_Y*g俎:o9wh(cu8ז3E^tWwys%c6=VVRwsIVw4weŎ{޾t[w7V>Iiv2'i>9@vq;nk~c p&Oj˃C`;1Ќ1r):th,lOcgPj['wfVKѿϻS|8*jRd3eb3vxccZ'L)kK9db*X!3%Ph[]+K*/|#Is$*RK@qQ̶9{D\bPgvq~50;[9yzv{cI.^*71WC;(*: }px2mҝwjrZnzƍ+i6`Kq2vtyuǙ&F[ζݼw盋{vqvh_vv[Ṻ ?t[p/]$fa9TW5Nw1 .aӻ܎iTO9/%# byJbhMS B])\1ǰ{_a4${xF=< nٻ*S }PxIBɾ>OJ')rRlCXj ,({K~NJ>qukB@%m QPHVr&tNVbyK\lpJ\@-$\mPC*zߍl"TPD8&GݝZnǘ;=n~r v7@Q+aiHuH*ДJ>H2)),%=8Eo$A!JN8`MDx*,#0 pd އv3g7~sAnq("ΈhxDĭX(1G, K RIBR%慲,1aACZjّ g"϶J  "IJ #pq^:{yɡ:;WVE!Wֶ862Ggkނ HV!5zŧgPuXȥҚ jJ;gfǏqA_VGp(Mw%O\7ǺV'uuA%/y{kчN(_WBcdYZE9(kuAEelƒ!.1$?$;?szž}6B{0% ~K(VJ \'Mhz>Y,M<Ă?/gY'd>+ {o`~<eq!*Cjۓ3RݫMV|u^~4ЗWdMpr/X_sɊ,|~y>[,%0Ol&Ξ؍N~]/oG`OV)NW{ ǻ)tHh~|of@C e1,%^>HYOVG7&2a&o>~bn<1:*CvFUDyi^)AGu'y|@p^i iTDF:ݴ](?9^ LCŤGƉ>-.k7ђ UL!i\*ל{B]v _N"tJֹe8eƃY^ h vOo9_8sO?ZHvnhQ..VHd5i%ke|ī߼\)b5؊aV|k~,XxTЍ/kv^e#_7ˋ3BOD҆qPu> kV/&.kÌր26 :sl?}뒻m#Q6൱rfԪmT~˝tq1?UuV>6Jګ]_ 2V|r3O{5΃fo_[#M%wwV^;d-{| W\C3E+x 0c4ϨEq r{4H>O 5,'>tAMefe7U٦^V>CaEc2ildVQjdMM5+٫ٓ?_V_*|GaVwdю\ _v,ksm[ &'b`Uv^=HF4A o讃stAE$}"$PrZpm(8ƬR֪Ͱ,:-C}'XcKHpc= 찧=J?=h-4}#wǫ1_՘U#s0&K#))QC@Lµ4ܝs×vA0RV›1I6cR8TޡD-S)2 벳Gã%oDd#,C2faӮ  MJÛTwJXU@w1t4hεzWSY5I6^#LMʙ q~e$U2oZ`"<%Cdc}-?W9.ֆqs٢*I'JdUĥ:c$pN2Ej-@Y 30aWVOSu9䢉} &g8:BgR|萚Y -M[Xac!8 V"kr " QkSl.'ف 2xVy1锴'{ІDAioe]I D.bF yXڲRȪBT[*o"!%GYTeķ`4.Z=f7d){K."Htv ,*0:Y-Y8V|T.EAh <(1e`yUcͭ)0h4y2Uq q8f,q]ѠN!)%2milE;`7.pWOb$ JT@68M  .W =XPPG/Br[`V3(_XY#+/%d0!ЮD(3!7AwbF00.`Ɛ)|k< P4VH(@ Lh4TRD&)dD>U[cT :j,b8b #|(  ITȥZccD1n,d Vt҂'J ڛOhhD\ ̑F+8ڲ>hjwWR}YRܖԈ yD{S@!ѾAQh8PZ4A e>A=5Ұ mJ"6 _})[4*X@>F,~IT16ȉd\9?XFH,4NA5V޵5m#뿢Se@CMqUve.35"i߷n397+[&AIk00 y% RĥG{@xz w#@€uI@02rw3"`9. DBK[na|^D0p 5vaZYlK.E]r@H"6ayv ₷- xi5C8/#Fn)puN]*ΔsC)TꅉƩZ⺪*Jr3L(e3cQ+(Wk׃ ,#}d s Kes_l[uVC`*Cr@ ֏ d Gi9L Rkd ƥ"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL be f DL  jX@ - RrFL $91 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@_/SPVgO =&H+ԃg&&*ONL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&W11rhO P +-g@_%* DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@z@S $Uf1Sq=M~fMW@J**.XH2_j2}3s)pg i#XsToVJ9y|e$,=uS#vFB+0]l$ {M*phv1;p`Hve0·a1QigUmV209+f'=1)W^}U4ٟ.6ij1t1-'Ww zOc8ypXڳ~:x7,ӕIjz7n/fse-Fܚ/aV=)D8W+pޔ!*PLJk \>YڭrDefiY]ZQVbq|ߧsbY3?gňĜͶKc@E?¿i/Y{}hv~8/Op"g]W-}\1|\]k4 `{vYzgkt+Ά>Z/-/V8YG>#@"X?ݒk>F{ Y,* D&*4*Ě9H(0ґ-VtwۧoMveLl uO 'ַMB_xY ŕ`7%&52;QiQMP oʬ8bs9,O+}QD5⮣qgQ Y1ܛX/EnC&)c-sNz̓fB)+0#IG`ԛ8U+4H;<ȜJլYl8q h?&ww hk '#5\C>ntIQhf,IWfYYaE&Ҥڄ9W`s\KXhsPtt_m\=I}?4:x)!2Υ+ Z.E)'V2,#Mm"I5?~nvO /|pd^ orhʿs.XB^C ZϏug{:] GCmB Zfh=]Zϫ;oy0-nj l;q_쾝Sc`oh.tw竀cY`6w⬷`=={h#V#VGl5ԴZ>Ɏkھ'Oh6gu޷axGnSl/[[C{2WԞō57߱ӕHs.Jsk LǬ85՘3- icUD⪈,LLx$Hgb!EHNEMT^1,Ks9UL%5,rgt:%&EXL ghS 8s}k{6)1CΛv糍鵯m +r-TP6٢HMͼt>.zs^9}*/3Y s[,&fPLq y`%eҢP^se,Q"qMrP).?ť_@zl-[{ܑӻAXL&v.q/RrљE;X0 ]1W ]m)6SeErg#I0ZmY1+b2͜25r%@aua83Ht2f. V\X*|0*n=90S{ǂ.b๏'E+Ǯ6f.n{CI>d|~ڴYڼK>wW0$o5noc%lUAQMn4w1M~ |Q^5zٯ[qǘ:W~Q+Ox{L$wYSd)bYUT#KPi2Y /lfUt(SG)z|ѣ"99g4f, M{-4nQJ" ʇxtJ/Tz mX/$~D7gbF' ղ& yOS1{xx5٬Wy6}aV3Yf"ِEE`R>4Y Ymx9ַr;!.aiO4;'s F\R4ҫ(5"Z4XC~nw xEɻax:;K=~޷6},݅~WRh><k+,pRAD[-g)}JscY[8aL =!p%C:p+"iJÏPFCckG(.8-\s]u ? ?=hYuuv@ߺ%Z fsGkx\A 0G0A+Ws8.s6%gG cx44=dy4cLd`JWxEl9M˃p+|~2A =[GZ[YyaΝLE!CcN z<m,i}oZX'PMhѡ>}B ~ը'ԇGn]N%ҿ1 gE*㝲`sa3ܖ[vκMNJ{277]ա;mVj97O{K0 ~>/&4!aK\]|.gp?Z4y&#̾Y\x/l8o8{VfG9.=]\`jO9|c7'k c`);惑5uZoSlN]}?z~WD;e,?*:G/WMOeK&ڷ6v<}NO TߩU6gQ!r#ꊠfP?ނ^bҚ C|g^Ħ+hV?q N$99ua &i5 ٠J=q~UU) ; /uʷ]Y0]gNNɏ o)=%6B]Fcv2=PPw=%x>8{閟t1v2Q-nLM}ۡ-;60m{D2kI#{yY.n٤ e>x[W%f_O"f0M/+EnkCC:nv˜e:냢%.FgB"34ڙ&c(_!E[?z@2Pir"g)Le{Ҙ(C2Q,,rpR!8\Jd.s>'Meۚ8nPe8>FmauD :;F ,*E DAe &CmJm!_PlǬ%O8Ilm; m[CB]ڭiǗβ_ -TD" NA(,FriW.yLi JM-6>8PLȐEQ!_($6j)"h|aZ~3Eq[ӏm=n#G{;gu蹒}!NGInkY!9 %OBs% i]ֆd@g#Qs# ECdmjnK=RW|:Ғmi/_Żq KKF&)w^dn' V@B(4SK/_| 3-ږdGpallj{ge/bC<g?>QXXYRRO]OڣG!%8E >ck כ>iON{W]AeYVRR+W e;>Nגؾ`+_\,+V\f /.P@ =NΆ A{n2zdZ˔BE^hqV:a }(:MmjOk}ۍdB` T!LPppkMµHXMOPC@LG0 wyWӖ="o4|Gv5m)X>c[[n#BjJ'"`3*, oXF'WqsLD&x& ظh@Kt5Ǜ*i2 8^~I_ ɕImtb݋20Nkavwdi)[L7] ēf^wzR4 {&;cbT,3~Kf_ξMx/烣Cs(!Kø5b%c + 1~xr s\x2F_i_&g6'^]Ϧ^/€k=:휞MޫtuaԬŸy~}{K`l]KֶjuKqlm3K^QIv2{ݫً]9z8ֶ*ݭr]kՊguGiXO˭/FX'#~ȫd}ɇ57]|7դMnMNW38#u<'o^M՛'\ؓW4cpZ˃X/6ʶ47kZZikLO=}j״{|h(8Wӫ^3eՌ&٬zŖN]]Rf?!KŎ&Z.+qz%S׽SB,@@K<ʐDz07gIq{W'94Q-h'0%SDCNJ;"C$ `>Nz ˫kv5;X׫ߟ}v-1÷~-IW+4ۃi=ҎxV=њ-Ѧ1G5`jIQ9Li۴ic"͍KR6mRI ^r˕) F'!{gRRIYNҦ#.y,2#:RqҸYN2ENVyL<֚8-k"eF$ e„51B`Ԍ't^>^c^H>ģ$: R8|֚̔d˲m=Jπ}lRwɆRRPE1 š+čNyJ)ݴ1ke1"{Fi HIZC'ݚmkO)G+֜{zad4LKcVs4sGjeCC.vVt " [[0UXFr'"DDjomm#W;AfUp8FÐLQj\ɉ¼ x-3Bgnb=8'<9]mmP.l'N8|wziʧ :L5nѽ-2{t&:#qȘV^Ed\_VC I @+pHae#nMx,1Xʶ֎.gqMD=+XL;kw LȌԦH-Iٖ7x:xHJJ\ t7*fEUe|qV7s5ăNFhE &OJ%{%~@{KtKO?~,v}x7j:_|WW_zqtrut0 t (#HFV8LdD2I]i^1}?3YHcsV}vp# Z7hyPzB;_BB^B'd.43/<¬t_d?L=%aT't5-8rd>>#{!򥇵@L[-xGGv8 tKIy ux38 =?Pk09d8q-*QRK|+zz<'yUسկ[k6E\cWtuJyG_,Ka-4i_Hrcm"rӨB,Be%aVd:뱳;qOGTRr=sg1r@t`3)u$(iJIrDU)BJZCY\6`ι5\;8'!ts_Xך qS~qG?~Ҵ&v$|"i`RJ6qud$Z|P~}V`Yb 57>4m J u]w]fEiG֧i>Mw-]ޟu|t14_!3-$K^mɢ <݋DZ9a'PlwR1& )3DNȃ3Ha}HmLExNuɰUaC~"Q*㘘|<ެ%`};̒{WHDG[*#JZ2> "yK 1tiq' xH:3MS=+C<ޛԤ/D9'`Y@EAĘ6`qȹyj /ZnǛ'FgaBRdIƤMs : et dhUti}5Ú{^+tv[Q82I;ږ lN,BVK5@(줛CЁf<pa cn˷YuӳOXrLJmfr{rQ<В/ۯS~ӳ~$%ZNOm>SOߕhq\5>\_ᬜK ROF/G󳥯̪]{B,Z.^{Y??tqWL_ߌ0"߽]>ocz{7Nq__4`pM^n|@({0p2H9m~\ʻvy6Yz?}Wgd O?K[R2o-@p{tKon8j4_훿ܘ~38[f}[@I^%!vڡ9y&G|o/m/JКa:)~?#?ϻ%ڿ{8  '=pv4:Ѷ."|ZԅP_ 5%Ӫ&AZGI/♗1sp޿AYhGZ' "m{ԥ_.xty?!^ TZH08{;km~<-w<O8B4zO2*uWvIt3[-C}Gg5A4θ~n4_@oivNsNSocy0cl;afF.ۮhfV6WN7y|B/NܱHqp`oyz/jBmI)lĂ4qvb&Ś !\4Tm( *FYω;<箨NWTgOxTLҨ] |7| /O휏tPo~tgi٦9?͓z0(FݽKw] MSs~S7tcnNoA ϳӮ*cŝC[wrl5cj=, :  6Oes Ԩh?J 7b[5av$G*A"x #=E3OnVw ۿQ/mΉ.1.5S1z:~^!8Gzw`iRd-<[(),n2;;fy}"8S[eAj|M_ t2:-RQ'T vFӂjޕzjz-=((uLA[c3ĩ TrZe-wV30WIP2Z=(+,Ǜ+2bQ=)թ^ G'b=*Ǝ&Q́ރ1mԍr4&ӔؖP\ ޑ.ĥ qCQ씰ѱ}XAbLF01MƨSA e> :\$G#wM NCR- "s&` rK5q) ow˖j$,H)MK|l'/S蜼J`.PILWj9H쐦)p2nS{Z#@kdށEW؈{{^^W "JY;D\VHk2rŬ4a\Z\uڵW@>[?{ aH!&{yYD Ҭ:4EcKa***f= lҀ#$}ּ|"GT@Jbȓ)nl ZOF' ABh ieėrŽZx*=~*znUf]\Dя v878E/?> aQ )UCS4*D8k I{|`rک}MZoFbs6~mb0syvd.*a $0d"^*a2ވ'T<.sWM^>%s}K.,s6VM <,>ryn3dB%/){BG^:eJ ȸ&v$yyg}Z2mg< C<_/80EG8gCU0`Q) kh L=|)=D/YMeY靱{fʭ0m} jJ)$" pYBI8Y"Xٚ9b[Its Dȅ`I*KWS hH^--o@Ys_M-mMRWe6,.A4L䧵(WU˿k*ӗC텗+^˕9-:!etɀK *n!0 28gY"eV1 C&/P :GK1]|iM쟏KjеeNptYC5(2Gef tyt$?HH]4*^wW}l'T)U"ɴd4KYS+6ӟ3 q5268i!"w>p;-t>393ѿ$|1XK6a 5 X6MDD g9( <D"+LD$}* 24}3O9h 59hɝ\&]C+::62!,пmߺϥ7 Š{烲Uvp~*eVG#$4}g?rRQYa=v%V7ޕqdۿIK ͬu~o*[fmPf0$f4zlkr6;rI_ qJU^ >Wޗư&Ū\OTb}X7])v~z}/_俯;׫^1e?~:s5E߻ޫ Z?>Y+f5ܞVZO| mVyM͇76mƀ8S|^3ч)Mț-s/m Uf~xyOhJrB3huyS FVќ~7dY#:;<y̓BUjiQսf|5Vˑϧg:._voհyE^o6;ZHqc"}x$[^uvXA5GS}m"5 'üyy[ wM*W5kBW'cVW}av|͏V7C&v̚aJ8;ݝSaT X:/m9EE^on詫(`qBm(`"LM^IܛweN0Co:$pљ7Ķ)`w׶ؽXO C'u7kM!T_ۋ_Ύ~|oհsnշbݫaZV {^\NnnGf/%`Ϫ{btLbk|emu=rST}D&P8Z%L*:gTP$!@b"ʶl{һ3 _̚]ѩ‰'aIL']D`FD40B ޤ@p >+vd}f84Sƀ /N)Z%ka1$U{gخݴ1yQ,Y!x;ǹpFNK`QKJ ;z4/Xhk=Z1RikiHHӡpB-kBRD$eEHZD.j[; SVH-6Ļ0y[bdj$pވd@ %`XFaV :"SL&ҮC18wx8vIڀOL#5>[๸|sxщP6c»VЬ-2N=:B ::_'."k{V`ף u:⭯%5$*j I֐m3# S{!wĔܰ,c$J S$& 1&JegXØqIbN(SF`e¯?rqh+oo;ywp1d7O%,}駙@!@&1> PET":# I@ŃާIpkw[*H\@D hf(ř@CL FrT@D7% zֲVۍx&-1gXʶ֎ew:QOOtuu,z7*']Ϭ?ʼnm2f0Q]-kyiY*T HkE()lxFΖlT/ċ6e Py"UB!6%.!bx !H{_̇31yA [&n\ ػOD'npc\:ݱ$_뿟]o{9t#SuPǍ07욋.)6xPj< .f}syk_PJ;rgo;m&f,|C y9e|&jlr݋NoLPޭΛ9 Š7gAW؝kvc'} T/}_ fE͋X0`mo~ўpRlzUk q6G4Z9yߓ<80{ěMe|z4EɒkX~S/1g]qTsZ3C X5ak@P'p:d}*k@LgLdڀvq*AdI&&G>M !'[cX ^S>kxOaeS,h*NY-.+%ލ񼊥@=Ρҟ{W] hƸ6EDaO>u~JX,9SiGדmgŀΠ|&v9Mdw=-.E#hP4F@(E#hP4F@(E#//E#hbkE#h E# hP4F@AGtȁP4F@(%t\(E#hP4FyP4F@(E#h,E#hk)\ աML>5C)M/C)=JaH,2ZX( EbH,"P$BX( OONP$+ ū)GlBX( EbH,|ЇnDЧq; G¶VQw\ ASY]ꀎapH3;rbЉ'f TdC\& LyMKT8""j@,$X^Hs8!c6vI:jG#u o<:E%8O` yj V;˟|Y}s(oco۸+-")x l>t}9W`c"q54AquTS᜻c -0vGau0`DL8ni&ZeNqQ ΅KieYyb"2䴳FFC4Rv'6=)ZK^":˱`/"Vqn83A?J뢣=Y*.n+gHFW# x#Cb#)k{4L }*rtpA,cޣHFΩhJI1kA:Jh= (dC< > 33!ʶ ?φaX'vUQ #AL%eۊ[(} 26[lYwȆջ贺5U,OiMyk'5C( $: OX NCiWQY@5-[3+7 p+ZnO)nޤHKv~!:靥D0NڂEs'z9j.69s`9\1!"y_#YlP}ȃ8N6] B_%!/S=r$qHJYCșfwtWUU;$fVbM}"<_xfTa1a 4-A"+"FP>io`Adxt~ =ڐGkx #&NI-s  d m#^FQߞn'ƽ;=K}x.F>Kw_9s'BRr[LP&(XbC&Ą̃=B`0@לc f ux%=JJ1#V6 #340eJr++c$wSd~@|]!{S봻TNva89p9'ރh.}}[ߧpo_K}͎7K?IjqRTtZnybҺP|l$ \f<ٴݦS/ ˧:n9@ G/|?;{׳z~{Ý]t<hy ~czAð]Uw5[e]op o?/.NzK`Z37Ώ9BHow8ofzMP9RɔE1"-ڑkH C& t:1N3St]$ SaPr~pòhXGAl!w`1D]+u ݽf Lo;]LU1;y=Qτv [WZF7liI)^BJ˜ipEYSK0})iSdņyHW/m%.Nx2)@vGG eG-Z%Tw@岇z),a*<).)\#h[T NDF,ME^/xXuIxWeOw>\*`ҘȰt^MOâGvyS%AuEG)\"[m~j+ԀގA#5F°wafKKk`5\>ksűyMz:- $L7 r>.w9[gݓq9kfc:ܲ]җK>Z-]7ZxǓygRn]ſEFҪD]>1ھVB dmޔiSjNnowvqMϭ!}gx~\LC<ηt\gU{[O]Vš7MW@Xmcwwߛv_"+{ev͟7BSb6漀WbyٽYGH(=wJ>a{p7-;ObK@|HĭIÃaV cJ#68R$#Y@{\TƒXe;2N͉*|QFXVoul4NﳫPdsh n^rszk꫹PD~g{NW$w=̫Y[޿!8#j[2T&XZF⻨҃ID&'?I1BcN5hq>,r-c1sA+Ӯ78Ŝ9s˭2MuT|^9Y&&VQRTڳT8@ň9+j$`g}q_՝['}Vu2}ֶvSL=I֏b|eZQR,C]>^";U$&9x4Ǡ8DOXYF]ޣ/(i ,hk&+ RUd*xus"Cd6eU7`IGwKdS=Yp;s ,&X@cC"G-4R >\) j`HjXX,@vGE3@qfAH$*Ű[)r= jolp2W}8<tvgUWT!:u+Ђ}ETJ32iSĘ71P2֢-brK%"32x5`]1Fَ*aaoP,d=c!;`YF328W1|sv=sѬW r3F+Gl+HLjb HVhi#0Ԭ2SdrLc(-Rd$WĦF4$gx !M* %р!c*/ȴ@6rRؽ-s_P7x''`SR@:&#>Htb4YNAi#88f!exю X$ 0>r8A6?{#g;N}?$ boPD=#8 7bqÆ#krXq` Yy FߴZ@BB2g($^2XG[% `'MFbR[ψ9"T[Yso:⒇,f}4TI=kC"V^Z%@pnS^\| \<;g<|`G#MPW6 &J=wGŞnG!gZx1-7z4 P:.w-AI,(2cD-5J $aiB4} ҷbp!oRMfU;"E Rr h'T^34Џ%HD0 AP$@2j^w_y, <a +yja݅1+ i6|HBRZq#i!g]jE+0ۜ[Ü+ŚP 1*- @%(_AzY0B=̉k-Qi*G#{6] Xш;*;bh ;GG Yg>ۥ_hLYgv1jHr*r~*OIMqR$q0SNA yZuQN[C {[D0 ΥS70Yl@x9#J`pWtb 48I֓ kx!CwVKgB1/Kcq|;odE*W>lbJ0ou%/'GKpBkXz@1H-kw"OVT5Lf7nU|-Uw~*oX/ be+4E /J꽶H͊rՅ)/vB]# 9GtQ0mF9!o0Ťyѳ՘|i9*AGmkgUHs`S^bVgjnxⷪaQproO^'/Og/^a^|uv :70. Ap$,Z8~ W?<䪯afh}>u·);ƽ+>ʖ-j)@f?/>(ÙK0r#m4&xdt%.|曟 日v:\EIՏZf]B3^]z}IȻ$~>FO'i` 31;=EυڡEVan%>I_z?qmrLSh>FFV"ӹgzȑ_i"m~10Xlfgd>Ek"KZIvϟb_$2eK~p"uSbU,~6*`{5QK)K1n_'˾A@6&v:rØշS Ю;~ a2xVSXՅFvu̺$&X1fYnɶCZ9iJDf^xvXLY';3HZa<)ކ%TR ᡸD'o̪iڵCNØ7 $ }0Z.RMR~(|F荨^UsRB ^X EmϚ;6hfnQr<Ԗ PuQgwC({{w9 5 𪗶{vj翿ǭ|=KVȶI^=tm'o}ه!fژ%L#znV釴Xr|1_.`ëA9kh+ {=6 +K•/Y44h܆1/OCgN !o.م~frg@.|wG$`u ȣ`Q8ԥ.Gm` cFI-7Ns'W3H0 + ޘX{!W=[]>)08}%>ER啵ҢBlgxy@YY'5UPwf҈skR[KY5g i:m0t c"(Ǡ1pHi@Iǽy\p1QKp3"ا1/%% XZ(wD-Ƒ̆!8[P[o; oNw;ߎ[^[e;j4jp=wCbpb)^]o^ `z1=N”S7?~k.F{&4"x~C\7/%guMf7KkFį?Ü.^Bwܼ?hR+@xeU`&Ls%۴(oAtՍ 1f 6$vt4'm /x7Nn؟݂ZyrvaFypsfm;}Qy~YQvM^ e}#im~k8:UY,O jG;#M!PjPW}j-dGNCUgR?>9NAk~o{g} >+;7A$)T47U5)NaxLu)vO -PυAf0)?,l=_}9$_亩eyy Baa=߶ >:HM!t~T9\k,\DF]RI&TR5 oUh]Q͝5wUWm hD(w:+%Ѹ&h8?}ݐp_lEuvPI xcb<2.k:Wg# 5h4(xz&7u!/r(NEtNCJ--s2wC@>.ςFi 0{us"CdD R:M#Jγdw ~l~?2PHN[ԄI86$2qRIf)S`&a:CDUPXDg'0a%r8{+%R.fgO)CR9f:ldWX+`dݟjlt*؂_PD8wB b)J32i)3j&qJFx(JH1\ >jK "32x5`r[llJ6[mdj Yf[:[|V3eB+[RM'kZ_Tm I4WL - s@J(3>(4֌Rm"#yȼShd Cr'ZK٤B:`^ |;t$)FN;Elp\CFDZg;`<-"{KuL"G|sP$Ti̵7*r*(Y1!fE;$$'QNp{&f{X9md"fZD""vq*W<~68)y+7=|VuR)'`Ao8XG!1}g  x҄Ij$f.q>PKy`'-GAu%E..vvqטYP$r XYd{i ^]܇]<{ktjUf{c=\bY%4X>$X}E?}s㞛ent#')qTR*jzE?ļR[T:_=`{y:>㽱sdXC7f,hT`(O/*c{L,c05%JĨٻ&lUJIVSݗb;+SI0`Wgs oS >^Ivt`n-ڝ @& vlF885u`0&c*P{3p4Dnܝz/0SM{؉x[N5mؗv[5Ck1m7:u5yo| ܛ=Q4F)Nh( sХ}{$\.%6!`BLBkZFe)A@ ,h8z|מκsMaMP՜ld6,^N=A4!T-_W}z7qIUtzxeYν@_yhY JbAŔ jQJX$Q4 K@ M&O"aD4_p!w$ ZMEVt*#R"%1v2J5NC!H&O)=*EN't< XC뵒A"nqFQi#0ЀPUJ+Nx:"1:DPLadO M%CH AkXmέaNbM`(sÖ] rv-!mtPk'4%* 8\hp++1zGe p~mt$p\eڹgU=)"_$Y[psuX87~1=b22.xϓNw@ ӊECR8~&) oQ3`)"KA)g&pjgi/ t^kW`wE-6NsTa}wa &Jdq|vR(\F82j!f>3Y>^,Գ3{|WrEYR55Д 4׼tb"k8IB*w s㩟ΪǓ{ nw՝WgI}r3Ŏ=5_EOyz1KfśٴÒG YL|X Κ, NV fmUNc=4}9xW̒'xp6<+<|P.TzYng?:^>e_Gx~3:{0Rі.No݅wy.4oݴ䪫iM+)u=W)r-^$evKr$ J?dof_E&YF=Re!1M!AL3 d~ZO2W?W{X;% Q?l^(MR&q#;ݡ$MCQA<A10T`26qgR;4*̭;; K.-E0N92]cNFl~&j)%,"Ftיl5uA]Øo# Цѝ-?0*wV@6da>|l %$˲I. vUз-}6EI}mw35`ӹ6_KZn-A{&}Q4Ҵ Hz\6cchΨL[道væі璳>,lzþ|i#D*f(Vs⭉L WF#R3RBFxr¦CJbcKԖiod(sʠO?eKoao]{?e{쥞e{kKpYX }w$TY^=NpizԌcۋ)LX)7m颐^mZT!Y &cO""2(,@d. H?Iu+keǠN2j)E_L͕vB$;W弯VsUm:ޚKI/}8g̸fQ)/?DLOC8;qT.|FFA˸4>\SA0!\Bx Vk'wiD0 7]Tz}5e!`סj[BVh%ڣHav!@a%|~}M24뷿f1UOlZݳ% iPz) ǒ^#xg`oMOZC y>o7 ?;RC`ǚ`e*" iHEZ-S7Q(DSUb vw&7(E F{e=* :0X;%"*r"1yCXcLo YdGdTaKHK"RkTK"EhL$xX;@g/71%6y$$L1dGKoA j%H%-q[z4ۦ>vmk=Zم&4&A͢4sBΕ-Oc@$ix`,(3xn23ŃH6 ]0yWqMj("YXg GpyMf83Œ*2&"wz\?aÐ9B}Z$}$K9} < Ugnб|;] oePpiGt\;~*%#.kh`K1OG 0Es1*7lhCj-Gmcx59%7KcZɕH!,bȐYmфh9I)d"XFt;YNILum9[j}*V<:W?M{̂f`DaX"gbh ڇ!Q"!@#`hYY3Dðqc";W\몽ޜu, JK]8/> țœGM߱_:mlݲ]veBk}q3N@~ `䞀 :H*@V&oM* Dy S#N޾ڝz4i~z}jD9-,r~oȾQaqEi= `^8gt`x sZDiKr_+L﬚=#"=vM'r`(ts! >(1Z2)sʕE,nP1G E6]G`Q0A[k%%46V:Պ*Έҁؙ+t%aR-=^TUPԣ̦AUYlb o_LHezPMB5,Etlrbm`ބaw\Erq>Jܪk! C)Te)}G@l?Z;׬M rh(H[lH*\l:ULyEy8e^xUS~'B8oQ">t_8#i0` s,Gxq*== v2fLR^{e,:u~? 允)7.ӰRneve8|;kH;|%gS4aj)#Eb7`U10 S ͜9s˭E:F jsj5Gf9]%11b6&8~q  ޮtHNk$-&@Gr&zҷrU>yZ6BYB .$իn)ep!L}ƽlIO&=F %8WR?W|.ʈHڥ:v4 g rt^PL5U:SzgJu4[*BtYfOO[ EJR4E^0832.׷N2*׮oX NWn #tHZ޵m2gY5|ج:o0H n-#`EɢOyvn`_M+]?<5ߦb1oQsUBeY`o<:TTPJ ELPdv:P e6%-#{`ݷq8O@jٯ>-RwmqIW=.+/7~5f;x2adsؔF60'MdMd.VEf圪 YvMv@oϷFk)E ֆsPYN7-b1&`-h:;F:r];Z;|i j+Ja8 p&kg 끸?H 1 ײw=|.wf,0kx30p'̍-Ռ iTUiԒϟ=„;R;Qlnc0j!SwUa}{D~CHaTDLGm?;9̵*@L3fkPZGuiE Ie@4Ֆ.c#rNc)5)أԁt@>+`-IsX?FJʆY Rқʃ:_|" b1:e6bk(|-E ـ ݑixu dF|V*M:``TNc (JG5lm bE F@n7 (ʡ7fAydnq8/baNc y FWkVw5TMX;M*4)"}Mܦ{g6>HiNch'~FhE$ؕӈ_t#f@57ki;1h aYo]D 1s;&E"Y ӄQ1uK$ᆩB ;|` j@Vd@z&+tpojb:S %a\!Ь3K* y'x)(`@SHPS* XbW嬅Wcܳ.׽6y(`Wz,IUBQt ʚ9̀(2l=x_ZU/pȀgF\ ߓ6ng/US J(NcFzPIk!䄄K.#f3a_}Ova}zD~pvZ5ڏfĄHyu3xHu@K Pd6Q t56bHh1L}qLΣ&ƺE ]Ao%J#Hʯ'T^SeR,zpK5732A 9`ɀQgk#Э<(so= c6c¬Hpk4=WwNu2yn&[^+I|wk]ɼ/1h/f9 (T*%B*q`(#GQp@ʀvo1aJp[SY)Ǝ n:bu5Xb3be0 9p| MV<\\=vvjc:& IFi5e@?xPZC"73 7o1YaX.(*+w] s#Yh-Ѫib;,3*55̬fz8PRj_Pޚ{%E;v hoV I$-hVs?r7ǥ Fnczj`*ivy8ۜ뙙dۮEPwS<pf,0zV61 n{OSpjq(XuuLk昴5g527l n0 >̘ݫҷfMj:=p7%<%*`OtT9P.Ϡ ڛHt%\!mEAv0 )aTPA>cX߼UɰlCVh7lE]1("^S:Ԇ&nIJ7yceU‘r'0- FZ*xTip^us[NSp?{m mT11mull"rH@:pYPAZ'.@5jnYl:LX \OH׶wϬu~{Pqab2|ݕ47לChvYqp$kmikMyA6P Wbd=lZ=?-H(A`pgsbt~V-jʠҘD|9RQ Hhw.P,;wQĵ-0 U&Tyיg'~b!wԢ%蕲 ^㱸؋[K"2Z'F5~h9~h:;=9ϸtm\Jne:ojZ4L7\T۔ho.|zJbʗ~>RMQ`_[볏J~6}h 'Ƣx5*\WB)R1)WV&܏"r|7&ЬV>&'dPrC,홌+˻f=}G` 1s;/Ic){I^^oL3 \ .7o q{[()빔{y ِ{lZ\R{-rJ99g򫪪Ӄ~O?X BG6俏CZ.`޲Bjux߷B32PY5&'];M!~*:7_@E5 ? I= /^H7"zx|e-ٛߞ^,N3׷5.qZAsϟT]ǫgG Ӆ[9h SnZ4xLD-O/9ǧyu%zĝ.\oT6f^omi qp#Wo0:ۡW;bxkڇߦzz͸xxHmpSζꂉ.f>T+q9Ra;{R}(}={vu]}m98(W|.;ƙ},hmXOdJۻ͢G}~X裖CjE_ӃiիM!5[W{xRʇVNH}h7qV|{.4l;˦I:{ަ:]ͩTbrQVߨs/bˋ+~1J_;?ז_;׾fGr7E;X-oWLwxG%꤅l %H##nEn ~ztpϏ|`\7<7wۮy.NZ|oPKɛ(.E; oX^=Lt7L\!oĕ?NIlyI]9qQ^XW9lbMV&৞\ͺгy^Q9 7~|m/}/\3HR>ԝ=ڐeS2 Q؃vYPZ-:FzP,4+P@i6sJȉ $'NNxvr~%9~?OjwؼM~|U%Nž}BjϵOO&՗ii>Z?^[FT_oxPa$7vߊh5CܲqZ=j,&ͱ:⃃c%x7_!vh&nvˇb fJ?77 LfӤSS>ļӉݛXM .SM~2%Yb&5wUJj?jBCa-7kDUSվVO*7ci.U0yoIa?6_o6-;;AW[YIxa BWkNLf{*yiy=~D~ضX'=|%)j Rs<_aGTl1ɦF6W{=Q/f4\ ~zfCw6p\WG-KxQ[3@#Q| "AK;Fx뼞E T~k(~MS0~ݎm?Vo|.6}ЎAƪo }mȇw8 k ;8$fDEg|Y.@cG BF{`@C:rjmr6b4 gLidT>HOZhe]p=lFywHJ?͐=" N.ss[[`LsC",L7 <Wh|84L #(`C5щsN?yD\r23ίB=#{Ccd /H)E4VüV$8DTG11oy O֏_n, 8_92i9 P͹1NG+\ ՑrM^޴/! vo1X=~0E(~uGIb_ ~Њp6r-jʧbG/[>69cT k3'~2hod:M+迍dtbˋA79<哋(DvZ^xB\8) oGM2 β/D\\. ;~^m^:+)*Wåڄ+R:fqW:-n{;`1am OG@o5E;L'ȱ1BWuHzRclGul=Þj*Y$?Vn8pZ{wGcX'7x0DMd#{dܨ5.x]pY1%85CpV|ijO~#-'BY0Q.jnOL/LIXVyJ_)>;1!| mpiqښ@49x4I!IcbsZN8NyKd:plSWtYT>bqP>h_uA' “Ǫ7|RM?]OLҙi(/v/|tYtQU$YQYP9+ANO S}H>$ω1:\B:+syl:pkD.$Xe4e*FiJ.x뜡2RD9Mn rLf`"|zBΚ5g;1Wowocr־=ܱ]/%mfL;ngὛm6+t*#!Vezvb?2۾)ӮU]N~dag߮6{PngJcvזj~n_M;nB>{ޕ^M١Uw~lNUޚ7O}L["c "lcШq:X !Aj>Au;W%Yʠ^yE|LLJڧ֥غwX毊UA=8ZZS 1PKwQO$A=xIˀ'U!2봇զ͉*U|gOtI0?ž`mw]~758܍;6B""KH!Up)`s*SNթXo]}lsJJM$r*khG%tJT@A!T nVf3p@ Fe2 3#D,Auw~]n49b;be442 HԜ➃ ٳC{0H'VX8 Plw!BRy NBP ^mGI97Pl.8Ma<.oq̉#88E< d˕'ƕWrm0ȵ $%0 wZP@ != &-Ói/8"FBx4zLF}?$ b18eaD="uKhl, .pg8ΩD.XaPBNM򠀥3#[FIHXВfU|vp+ٍwW_C\Nj0.{\ܞ1#Q,QƲD#`y"(%ʘNA(Rqx2xXJ;C] x+MlCۜzW~TXޏiӠ7~ =sp WVJ7hl_+@_%<J1^Mtga1`)kb8mBm80(3Y`jy][y.ɇ~rjc2lt#OhuފSMmkzgrA7NM! iB64*⒒:FɏrPtHL(Em18(pgՎs%+!)JgOa'sj+ +$C4&{aNE|->s~k9#PkYk⤳lrݲG9X*g wj-Q$YF!d}$w9+6iMcܬJ\ϔL)ь*)M\xS#"ZQrG;kIP(T8 i_m?%QM.TVshZ|(im`"qa?EUv?RS> If o[gKd)F IF*G"RҞ:N",F.<@:,߉Vӑ"H4$2T&+83 #( L+/|0aӈh2sjr\)ٍ/ya6冶dp x"|xСyg 3QB9ag7RI)gv6nt6#^ő$Z@J8j=!f> .'J Kq=vX_5)Y|Je|?A, `T.%oۿhtqq>9_TF`|>=ڶ>9)5wsUv@Z`yڃz^^?TK}Qwl0Ac8WurЯ+ $f/Veǎ[ N:b󝮻núocHmV8CO]yoG*(yH}THgc 1yGb}J\SBRo%:GeOawtW.+|8IZ۫`:^׎ՊzGXa8 1ɮ\ͫ k]F5ߦUrzӫϟ}o~[.g={oi٢/0I[CV$z+ 8|宍muMͻ.g~Q6C[٢Fž0o>|;ơ}I6Vbl91I+U Vq"і#̚h,UdAlD%p3λyB]ՓrPG4 z= PO]&p= eJ^"J,d8;Tp9(吣h0L2owz}8E^pϧM2Ff^R3( z@"g9d-SdlKQZ#dH)Zft@Z듶Lf96ƀ/_BQm̮4uo&[Ƀn_mT6 {U^x\?N`x8WոVOi,…?8+^ifjŭδjBVf~/}H{lCZMɰӫL=ptYw?`6|Ӟo'\R/(§>~*DJaLح~cFLF87F2}T͟U{hTJXDSBmN$@λj.nZ9G~`ɐu%XApPh_5_Qo*_K۫?CZ|y$[,T+|S5`g[0+uUl&1ƛLc26sl ܅!HѿlYjQ̫p\k[R歚XYDdI3k X!V^_.]xHzR&ͪ)˽Z =%4JHy+Co-4wG1`ow4f@^mVCN|Nb]c̒ y\gGSdeY>BgCmυ.IK:bI8G*ڈiYW{ JFr9"t##U;j.v[pιrYg#sBJ&%Aӭb;Q-Nu~gjTӛIg<2N#l<RQyR_S]Kq$u8o饗[ '4cV3jIgh0r.uP={,GgS?^z6 zcoǝGbãтTǽ8*O񟷎xW \ՒՂ?8 JPOF|kVݮuӶX*HbFAN%ԭ&At$`;Y̲ r.uj k<Ҥ+o=;", EԴEUݯ1F LNoN2ARbP;H`jk]ӉW\HI==Z㲠 | L+ lġNW:%l79f'D $g牔/n̽*:HDF%XZ斍sݩ\$# /j_43ޖ_]J v\{ރ}4/ {i^ULYlO[u$!Xc'Eb5UN:Hz]c; ;cDUP^:^<`()('$g N89ķ(SG ֪/t:\,ĺ$ߚs%7!{0t4ցJ k1W]HW>MF\eoCDIG" ]!l99nN$-9En-r.83p8l,=ZT T5:EK^WtiZ*~pj插T_} # ڙkQ,SGii=ykr1K`F&SҌPG*Jq!rAKe|IJ4l*h Yxoaa29cO) p]d\O0=OZ'fࣃ1Q)djZ;˭P6֐]] P3SD[Yr|`SM*^y&+>opYdɳː]^1ő`7> c &)[yHi.L3ͦIZ _z<߻Hs1cC:xǙEց#v i2;HcAxQe蠓w"Z-N0tN)D9ŜyIE&rLo0EɄYH!{.T+>N5IJ2cBarVJs[kdqq3Xk(ǖDO ĐH![fgڜ0z,H.y vp4ُY+z:IDU`<~eDaCP,iB!)/BBK g;VLZ KZ%8vXrOc?(Vbc"6vw؟fTcFQ}3{jdvhXJ^MM}2{!`ll#* )YCOhz] ,Z7-(DAU W6q9wN]t:J'@|wԣQ-H矾ߓ%DZ9o$lͿ9wUfۂmEOJ0 {2|RdW19p#8__'Ë_NjM7DDvNZ-+|m ? 6 qY ͭ^3 wnDIo XKt@'_'tpkJ>pi,b?NO-S woq\zhP @1Ia޵q$2P`~za9@wd`7Ogjʯ"% IQMi uTUU]`b|jH@׮A[ݳEnoc\d9?*Fj:SV`.>m4}_oMoӥm~/oJ{$Ah'+zU޽c7GٰinPeϞ<=h0ivLӬ^O4vL] ]A({o5k\gwO:c9왠;&ɏf{hzzW\36yoh:6M;D6fD3~:a;ģyeDq|8pVc, +|"dB0VLL:n"EkF:bԛ=//vZ /@>q0&=2BjkB$IΡbbx$q.YPJi_Xw1Xrz h2/as|2 ½U/?3<9ӜRt]Q:-%{5 I8YsgEm5W"w% !KPPgOK y:+syhl: Qh"B$ME(Ebhz%8uP#eQj5` *,la`_B1=B~M}2\Yсfz\K7pGҷ. }͘N5Cvፇ_L1t+I2u@hn ϧV|Ifya.jy|{s w\'CM@x3xN_3#ǽ OݶMYÚǟݺ|1G]i1PSC[Awr;d3 \͟Eu7z +CmJ.rRwnSjiIux]*a=֔z]ІH \T' `zEWEsY6Jу %&1S%tJTAsC`Vg=gr @\ OP0{G8oCf^/]7]Ez8ņZa{ұ@Sy-F\QV 5ALWz^xV+gY8 KBSN Ft:xJS1qKZb3c)?xQc'|p͗sR0*ah1</e b.hń,-#*+ep`D HG4gA'TSC!Q hͩt"895.%xy7jk/ɶpݕ :{|gUVޠ |!d.T8EVL6ri.M9زU-V gvd:$ UOkB2[%nU JMJkb׌J1]X3Յ.]1hUF= | 7OMw7v8_kl'Y \mIHNeUAjNnCОj8¶* 7>֐8Ґ=!T I3/ж#ڤl.95‚/Zw| eNd'Jppxɖ+$+ &XA%QP*ЇIHeӂrRVP%Q<#(ºD 'bׇQ?hbF5ֈY"jz핯lUazWw ܑMHdT5%)rZ"@ $A/B/>,C]XnYFq*`qE% ;D?>S6客gf}Z m p[j_ޠt{aOLvyJVKg}}q9x\`~|9B3AJsH<&@M p&eF2Y"!kk1 ]%ǐ}K(sK=}\^s1'ٵ.k(`s*E 0)!$8 FX\ h\zV|&y9he-U`|"eFAmi率$q-yoZjUo"\ujzZu)l]:fru9j:nxUoôkեШJ*NBs[\ռ^;2e5U5(2|p,Z#q?Ε\R#E+=Ą+J9)lKvh%-zwFL4 lL X`ND)!GE0KJ1*&H~  E>tfB)jc\l6;v"JWQDS0$>(FΆ&\MN\leu-}ړ+ߕFf7++ K%J=eiD{|MSԊJ s%j-Q$YF'BH&Xb<ᖾFp+xg-l۳.2xddJ H+*)M  TDD%,McPȸq; +{na݄)jpKe5(PF5~95Nk L"R휉 ៍"YؿM If o[gR$KI{xGPAx~U*K+L؎H H4$k.U!Q' qp ]qXs93n~:6cCA] s30UWހ8׭UeDmR}ܧ>ߔ>~*W +' g4U?M渙"L.`:'7 ֭;:}9)ZW͞8F,Tw!TU&U•}=m:^^:_,~UF`p26~{pec]_^q .BkzW-=={'MwWop{n?Fi88=.8HF?]MZ3;/0^sBf*Q=8t7 ƐigGVG U;x4-ts;|0YpFYg7kjM>3 d/}AݕH?ǴwwtX;T"Ol0.W~pqxˏ?_~wᗷoo]W`}Cd$ 7i㩕ХfSJ|邯׀(=VU?]}n2h|ۺEFvDxAWVףPٟ (ݑȠtޥyyhUF ><&R.Eʮp﷓$n'IQs4tq @) (<!2DC8MS`'}jLխ.=/k(H5Z[!%S39m#4.idRJ)ùNsj!_'JcA!%XmBΞgq{cYn#v'`X 8ɕIF)w5<66$\Ccl qV^TeFg7s}A"uZ6DBm.g;c$:@ܣpT!NC2̓{t.[U{8ܻN Ri=ҒAn@o]'tK&)\CABW5j( yPGdϊS{%-_Q 4@5g*ENl+=myɭR1G}z$,9ؒVPqC}$&aC10ZVyo>F 6W1hd ,CZԄ)$).W~ $A鵅潩Wmo&[y+8ۜUux"M}ut}Uգy}4UvbGٗ}]5#XƝ7w¸j1^T4ţ6o>37eJwYx8g]F7ȟt 7t7݆&};GeEz-_; X"mְ1C!YvN]ktٵmS\+rrZDہC'@P5q;r뮲dڜ,Ku|q4էj[VtNj9*^@|Z_;6aK8`U]uTxˠU ܘ7LY< 9j< HC\0T sPT c'5J]I'*6Ss,}2]G̠(N> w::xpjT<3N4+!o_z]64^ͪgExЯ&{1:%4q>וBcيˏ>+`{o;wφ@Yg(FCNoɝbmt*D&P8Z%L*:gT.j 6IIBpD'(6g͹kut ?{̚]ѨBEJXhrQUKJj@QPC)u4 8>+vd}Qsi^".T)Z%ka1$U@X}reڸ'iE GdYTBGi1ȱa0>ePwͣp~-ihLvCK=:28:~SF9ӶR JJ_Ov#Co})-!QQjH4ԾRk0]<b/"J S$& 8LegXØqI\A(SF`eJ>p(f`;o[t~wvbEe7ܥ|f]^YB4MbmDtF6"5M%OV30U!rhf(ř@FL ːGrT@D3% z076xv,:3g”ٕm)O5buCF"к=jozfellׇ]%`/>PfL6g0QڮTOky.6S*T HK!ߊ1Bw[X~m87#%!ED6ClzbSCX8OvڗOvz}.$*G20j<&*.FdU';kHP䔣SaFw~~zj ([|oȾRSv"~^i8/vG7`%,x#Ks4K\c%I80f#1"*b|~Ji&D)Q{/$e yʙST2q.@S7$nIψq9BS\ f#.}񲇔yzVʧ$po>U~B+ y c4(C!CcLC=-f/r~E5({!TߺnHHTI[f y(p 1cUL-,Ad,G+:iYF*A 6TQ: H MR@+o0WGt2jLdd`૔ZE5<+z[Vs"r2N EV^YSݔzPW9ꆜ7zc5_eT$PdB JB@PO\/~ |.fHgL v^=Npw6vr9cpu6:]=mi=XN d8HUao*<*JM5z__~u\ԧ{;iZQbwtt8n̑N{AۢނYmCw*(vP :.~=rQ\]ICh\D^ o7L tL拆 K  n%VQsr7{O-IԕqzB"2&1 qpq}韜^=XF`|m)SdR~,~8j8NU"Aq}y>?m7Ս‹ip`0aB'[ڭz^InЯ8ĶcgY֓=Xɲnn Y4|D& ӭd`Ŭɻ_MlK{e^G/ץԻgAGX {V4 z1W't`FU~Jꇵ:·N?Ջo^?>?7_}yʉ44C7r̲٦i.|:gi KnL2"e *?9ב)1,2fwn[M?yd6~PU d{f1@sw)Q'Q2RQ߂aMq'%r+摥!9wM3M˦;l;g}3Wbt*-3$⏅4rjJ^Ƞ9l7H끐"J9lY.Qh@) eLK^j0 C(Q(1Vt5(TS,e &P2h&Q9R1G} )zyGGMQ?}:va?"t^;ӂ]yK1'3~:lWlG]lo#namO)~]Qac/{VU{WғNb/(+@Q~g3 0Wdծb̐|u=,V1W!ʝ[@= '\>)pw.`I\Jv߻g;Nyt/V-scv|3gg4䈫4"AqP*Aqq.`t}F>RfR !) c3F qQF3*Q';Yc,i1H_̐"%IICǀmb٬{-^~J[cE+9.|y:cwYM_bێ~{nXNP4`# ӊV{͢TG2X}݅=<)ҁ~UGL*qg@ b<K{.heL`YijkFRuR.5Ͻ42/Q")Fg),#[AZ)|2_!xcFS Dʁyvo3mYeKsUM>fP9 "pp`LsM;Z Y=hmb]p`uN1ܺx;u.5xȞ\|_ñGG/lMW訥UXtƉhQ+b: bBU@fȆ8KK#3 HQDHv>u"dW^YED);d҄j[rn}~)䵕!:)v( 5 ) f|L2*OΈL*P\c9Zft69AgXަ;a0_6vau~ m2gTG R`"(}pnLC*:eY,Y[`e>Ip%GAE3 T039x̅;0e ekA{ R|3.毚ua *:KT|/'\p[#3 0nUODi6C(@`x Kg { #tܒ'Md+69H)|Jg)=bR 0pZhNLEy =o.̣U(pAdmٗ_(Ce%9LA*IU y# k"=>Ϫzg{Z<∪}QuJ )]Bp͎NR.^%8 %Pe$'J*; _E4?u4?Y4,'eMdDUw^Z$ͤ 8> 9G21c:O-ۡ{i!] ;K{5ns<'LFqT.mBKȪ^!jJSJW}ϼ0o.E4um>t^oZ쏈@a[m߼~QYyE=sF.iF`4Wh5QK*M'yNh ԃ>GesDtăEWim+>E.<6k^sm!d\邰.EQ*6C ; @Nr5]K7#H;qgN`%ّt2zuuIiB=y!JΡO%I~z7=[yxǗ:H)bt!o⁑{ӧ|I?~)F}v1g&J܄PQ |^K[(hE F:Yb>2l`&bBFg}-j4E%rP3rH-K$<nﮥ'S9JwutopbzBxU ].]OqXc)$.NJ4(alc{{0y,XlW.pDd:o8s ,9CYQ3-Y!c{,ӷfI< {~T @ )X' VsdTKְ$0nԕF4'>p킂&{U2u ]pI&pcA Q_]P"g?oYog"\{իOpߡW i$*VkG;E"+܁hMBb@GNXy)o-YTkxڣȞ%!Ʋ)R>0Q`5(*:D0à&p:'٩+{ym >z2SS)8U_"˜9Ì`K d1%Hxž7 6<7b;֮|y܌c{FiBrLmm!T%侪58k2i4IV DѺཤY^GieIb <9`7:_%~w[bIR+[j!ycTrAƁCv/5j";Ey zpPr{  *Tsۜd<Ĝ3 ~* )d\0}n>^lɇ* N>V{&ᅤ"KxmO,xQ@ Er`[ 5Y3?PDBE/x<*UIkUQH!&h=IB(<ȹOATr<]u 0*cA KW8\Q(9nse\`WPq6/p=:ͅθ ;ܟW̺ -ZxPp(`hQE]kvCe9É/u! 7N7ȼ$.޸'$Km_e pd)čX*P4Dx9rT//ƟRھQڞX^r_4a3)(&1ew3Z ,]}c;.D֢U+7:Sa?D-{|A-7ͰoՔKRIad:@vD@j *IkJP}mW34j)Hin rj1>owNX^nQcV?w|37ԌP??X6 |P"l}$>r;/q(.㯫Ǧ8$ k}7&I.~t2[\JJ%nIq:"\_,ߠ ;:~7~ՃHD9vf-;a/uO30yܨi4߻r+Kޥ;,R[KH7d"w}^"0p? $AKE.Q]EC>y->w25iWL#$:FU 8K(6/(cKIz #t%t=XwERJ6{s']n9셠fI !1َjF8Pd8&Ƽ͌l̫)=r_vRo~{[˚_pjzg2eK/sY>=C5E[!D S4 ̃Y.sB2Id3*+;:%j/fWid/įDi1\~~ eQ/ل{j>G DUf |6C+m| rrai>,adD'מ ͖d TTDWEj`ys=!\D(}ⰀɊ ">Yl⪊GMy{1Rl/*/:;_J&CSqsn)uNKKoxȝՑ:2uEodp7Uv/e4Gw[>ٺ{v}]`]W,g^{?*-E7=m`f=7E>^SWl//8n;Թ޽.LG}SCU{V[_(P33GԎhxPRaRHHK>O ?,~XI 2@-[XdzL60UzGz 8%ZZ( 2T!p%merf,eJ k988uǽ^-L:AbeoR[Z,sAzͲNc\Uz Ws3RO` 8A[#ۺ.fF$3zgϢrF{#$띱NlQru Y` DbZ-v̺\I3 Ec.pnaO *H֨KY-u-U1")DiX ]\! 0r[0F,pmCf*.IC&d%iАIK{ 5GQtUbj4akùY6J,cQ"[u%nY"DJkX:ĸDs\}_SH sM:LC29c&0$6"9o@,I i.cqpkY"6%!RYF:ۢuiY.\krZ΋m`U2j\fqR#ۢumYDXc*H:'#y?Z ݗlmef؂ۣG,UdMSGS[c,} $zl'+5<^pu+x/#$`F,1̄>v \8-- [3sZ4y-s9%}v>Η(?Sۮ'gC=7=2!ZCN"/nq0 >L\Fަ6jIS;\h*L* _QB0'n(.Hoy3F|eL:\F>V> 8qWd^^z>~郠Ť&4&ef5܏чnSqwU-d-ȮHZ :͞[xRd5ҫH6z .US6[MzdͲՋ7%JOI:Ԓ|r{]t6TG\s׵*?Ϟ@ҸjN:];r]܅L'e9@RRTk3]%cڝW| ]#;]Wmzj3#UӬOMU Ʋ9=P=!VZ^ټ' y@ u{/&_B.Pw`m&͖*X0 [˓|b]{cs E]ֹ.DJ ȃʂ2j-!٬k4ŌֲuLO*:wip- D)-֤Z^GLWl8(5wfߔ(5Dzl5ckfVXrVov >%T19 TxB6:3cbe*IUOCe2{D.K &!XI[oBMUpC2XjӣhcKӍ Ÿ˕ kROjj-Ys-tM]&x:0òlHFP@1l,d-:_uPɨ_ضa2wnʠSF0FAYcdC SN*B3sC5Fnnˏ A`'F,fͬ?g*Qx &dȭ%8X*qnCp yT%G!bnM5HQiHCj+PJ(YI$ rGC@喑erH&Q{R> "sLUyDC]n!sH!fI̕PjeoZ~wRT }nmڎ\cz '^䀘~9-րWXH|$Myz^|P* 1:+QߓٝXb὾4nL/s~$70y@Yc[7lm^#vGǻjHQvǽQ' u7k\0Χ5?Oh8,3(51չL&Xژ~yx~2>^xήyFGS{݃{w^ $Df_Nƍ‹CߟIƖ4]ٓTWtǖvH(o *ً=ꎯSK{^g\ҹZ0Qq?$dD+Ws`wx5xj\aMSܨ&mX?]!λ|㫗o/Wi١70X>/Vp!z,R?e宍muMͺ+o欧/~6C[ټF0ꗓO(@ϳK8혤w *⺟4 hr0UqD3\slQjD%h+?ttcњzG=1$A"F+Iq(}R2s"PeNJ;"CǀU0|'=a&{Dۆqt;k9L]V"(CK9luZ9xcEk6v3Ά:{_tqVm]qh{}?;pwٍxPi6eU{vQx.(i%8=6Zc&x\!zxL*oWٳB!9׎qk dln-E{9̥t,eN`4I S;!5KOlbTZ( Ap FbUGg(%zqL9˘f3;dQ ]~ܖAn,xAĒꠍ1!cǰ % #O- qbWz=lϖ#E]z2Ѝ. n9zFfb "OLJƒTP؅bhk@6tb<5=a s俳mnیrWX?mڙWd ڛח yHɐc: )S %le$e#9!T hU(`nIq!k@R#VARuzjLϕv @T_Kn ▿l'k[Cp_8&}dw4ZъzÖρܜ`dݖ+6s.o[HHK;tRO]r%<(-4HSC &$G#> e֜Gk8 5)$J%RYJňu.7 y":gYvK"3`DБ-t<&rB)uBF2l$x˶ֆB[DK d9im@m0!g2:md~4I;ƽ=0u&[QUJ;?:jtB@˦Pձ9=,rKFUj?VbN̯܅WiPz/x}ӝ m|7}NN~v ?afǃ^7.!L$ֳU,g1}\鳝^X Mݳ<ٍ?>"%`0YV1#&EN"F#ܪfJe.}*aM :6*QvV;N24Z_% p;T~}09^~'݉.# ǯB\J5"Ai{P _آhŀQg_CWfr1fJ΀+c3f.QJ "qg1@\KfZOB7"B;,CwVWHz_A_yWVPO rA/D ͺ2D}3}hq3>6/?W'p[ 5p,qEx-o(c0-.\g!8 O1,fX؞Bwy$OgIi% 6bڄ5GW{5 aR8`;tWkV1yelfKZ 㜴6%0іbM>;grko3yƽŲ@2EI;/ xm)`VC0+-mmzsKw>m̓>=јcֲ{sB݋jeCb8dB)sLJZ:"䈴!ުv.N-8\[:@Y'Tn!Ey%LRHEcFn`ݽy=pV\7HxGt5if~JvD&E([tМ=n[t8H3Gut4دcmK芌8ku}5\Vf[{J(zj?| pߊoNV;VV΢Q6$ ޵$ȧdG!..@EnEW?$)GRڋS=$%RKRS1`{0쩙[@@uftt 0+Qz?tBT^WŎGb ~Ο6ތYmk_gnݕU8$dIn@%t1 P- :Jz+D]071IEKX8#"FB+/GHB *7ka?5㙹p ,l[xez$t) iWͺ߰]CueT|/佺zml qk࿃F:xz/Kqe<5oB[Au^CG)Q|/I錋2$bN&2QM& bA:EԱr\e&Cc虭L}sa[P劋bOO>-xo54Yt嗤*II ČrFǠ9y` xR֫_j# ۈ S^Hej:}gbgXm%VT윅"J7- GDص o^H 8?]?3mK40 ZM4CpTO (_ԁ!NrD^6GDy16bYpIkC|9lQsbːJq!A:"ˤOtf `#w_a\U#g?0+@G͔VxV*^I'Sz.6p4)}޴͛`}^ض0s9} ĩWA[SHH͒J5pxL' "N"ģ.:1sPz^<&(/$C w:o> &2D칳yo2gZ )U^C=z͝ ޺3փ1'LR(CTDP2؈i9 \Y;'1􄘾euhg5CCz(q?|Y /⅁OA(|J7.U( IҼL*i)1@$0Zwbb1B}RR=EL7(Eϟbj0Iz߻ jXmXaN.0,ُI$&b_}ۦfLl4Im 򖉀QH9 oAv%WHY\BHwZ{sӅPr  DW}B {˦+IއK vo\4^f7 sOFQ /Ozx4W)ݒ?x'8~7'}?nð?#Ne۹o@?H~QEqay ?-k%=[rbw'}w:nD޾-p$g钮Xo/\MKq%*| 4nw5nq>f)L7FL|m6.wG\ wh*r[7."/gk}Ԍr 6-T2u>m4_GP?nvŏ{/9d_q fDe\Řj؟ML Y(;(Z(zݳӰtϜc3xd°`d Zڿi`yhL-gEtɲ$YrLFRB֯.f"yo$5'bԧ';d^P%Z o@v_,m̊,#^,@6GHF hg@̑m]'C>oN?VXhOu3ߣ2+“Uo=˙~i%=B׫]>G_xkdIjЫƃr䖫.K+s Y-מHIGrFf ѐtrct҇ORlӖ7g6+I~դS* ,\I)"p1xD3K.>w;>lS!!J 1ؒ axPYYZ>ZRɨAL:kZ+>{z°a|bѸON3l: 7԰6HwF?+SϠ/8k|c4FC)BVHu>y(-Obsl @`Gx-󺉽[Ef55>(Oӣ*|4C{Zu}]pg_<$H7wd$%?B3qv+Vwlu^^/}]n;h]=}J滕kyʏh=3l ٬k-_ay3Z/^o[9GooOuоI--a~(Vˏy;~TvjOoʕmt=έT/,cʣA* -}f<: P9t *>Ny8s *tMzjz#MA׋2Z[ SxǓD$>ٜ,ǸV9kdz"ں^t %IqHuo(Bfެ]?xm(z/Vz%9ޟ&_od6 :Հq&Dfֺr'C. fs:[&XV:+I"!9&J}u\#a)y .pT;@%b`V{)R2WXF!%d[:.Kkǎ^ bTvup#tx*p)W=A4WYW{i.k[X%%2hJv) )),y~MɡWwB՝z"ARor`2%IfZs%Lʀ!Ҙ Х[<i/e8X#مtx9"! Đ'-l̨Ϩm&Kn)dW['ũ$KxxɫjWU*I,RA-cJ;-0Ř \Z$.YY^9qnOmM;#YrA'b31;'VIQs-=+TmXV& B, GPz}3*xKӍ_a6n F0>F-RS,Q;7Q1jhdcDw %,7*[Zh(ʞIDA@Ħ&h@ Gt;f].Ld`le]GøbԮ:^{Nj#8WňDQF `X1, FriW.z?9FTUfhҐ r -:dHD%= #(:1fY :p9a/Wx*xDԕ%$b'wZu?{qm /AC{mF7hh3V"8)ڇ$:"-Rc #}fp֌ڢ+]hƹtЬƖdfI 1D qF؝z:LxĽ]s%KSȐ1َ?j \ AĦ~ٿ[vw0˜dxzi oiWFe]3:ݴA3kK;j\*:j)&CoCJ]-l3sf&Υ(׉vE(Q?iy\cY-5L9]{TY"iY\ѦѩV!.9y|=~`W|,oh<޻mehͱϵ?{zigCŜ>|η\ڼ;[M%wΗǧ/^< ?_&zt`p_ه ٸ:b*krsgW{u@.)Rf|>BvI`c{@ :ϿvɻJey2 r~hi~ngU?Wd'=%Za_\Nt wZņ5o8NR|+8if5RqSש5`\SUoGb\H.aޮ޷hq0pv>tZM[:s^JI2vg&rlICIZ+z5YYU.GMfdq+2iBOWxh]g;n6 y*|/:5wIek{ $GjQvWig9&UAleViFVuJwfUNд1{&h-Rm*SUV*MTn{} c-3Vp}m~ ph_-6D.cP\Rw+֌نL6omҤ&Dcri[Y ftcWId38fEE^b6ZKNI8_–HfV_M:Ȳz$Y Ay<lsB6J#TEv̹0#Gw!UV{l~%BsJqG,;I(%hO0#C*R81#3P2d]iMob|jqs4IH/ZƩJj[+s$+z[tz9)!K!lkuAh]R"wLMZ%1¯PZ)|!4mG;eTCI@Dƃ_dH,&#զlR\=H@`1zҺXt):Q} ]&|J'{XMI1!Ku4X:C 0iJ!B Q=5 }m.5ˎ0H%GxPǂ ָhSk|FmlJtv ,*60:-;q9x9GDiDM2MXrպ*5%n2x&Zn@kBjclu+y,X āNSZ6Xa=Ҵ9;v%RSN-ljU!Sbf 2b!]Lht( C),k:@=--V0!ЮpYV Zq3L @ 2X "Y("3m؍ttg-JQC(]e֜1 q04Ǝ\)!bD*pPRH 3k'D1G90i+XGeұK"\VA@9%pa7,T RI[v= Vkd5{RRR}JuNLcrtWDlZ42J-J)MZ neXE]ЈCk>84i@gmܒeICth/Z1#/Ef̺(NcD-pB1M@ߛ!wv 1¼a-?nՂQe]@!0h@ƛԁq@xc#X YiV2Mt%CRIUFB9L hF/y3xoD>"COXV91/ ( XB(;KSvhC*Kt]/sgmEe`7PJNZ%l-ѧEk@.Ii:B 9^AB ;!3ڄ`ctGJ߳N:jNa¦3^LrrTC%#qT{Na>XoCnҜM5墁Ua4D v-:flN.$C֛^pzꢀ%a $Ti`/tn#з+ygjp N}UQ~ԋB??˵y8bJDJDs77ާ9dgQ9waZU 4oC4'g!2ghcf ߫>A8Խ9T DB7m;9(Z@@F' N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@t@FֲBќ hAQNO ˖@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; NO | 19?' ;;;>!'Դsb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v}Np Qq!\:@2@H`'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vqo[Z?g{/i)=]]\^PxZ=]M __ZU)>-4׼x3dɂrxɜr~\^U ,giv~ѥ.!ɬ#~PN5ߔ4ˀ[OwY[Ypq5N:89O,]z9r<=%5Jk3OҪIσ?f4ݩwLHQKײ z=-R[j#t5aVM+g哗G4wqx~6O;K;y]Bƚ+rNV1)3o "xy|lj|O^N٫_atc/t9htޢfD126*d.Л+ɿ33֣`'Tdn6۟EkEK*ߡfV`3R m=灦P<{f`>[5b=x W "0|X:e6Pj?r6G[0uo[6MppP}kQ:ٱ \O cȟ_.5G'u=- JWRwuPmHӵ? 6^kE_QͫǍI!wn_Xv݇_=ZkYf'bwKjiRKCy4vt"YU|HVS˔3@S35bw U?`S%6Qτ&ތ^ʞFeGUEiA>M'xӎ 5˓q.0pLf@}bұ#*(`j-HW^M0&6'"A?{MSӧΏ)X%0/W->mLTq;kA-|2m8aao|]l.˃ E6]]K^Qa ce˲oɽ`S&YhA8WkKwkYO)zF`=|Ӌ:(UT>`L巇zJyy5|R~Rl2 >}QcpGpR6Z8QR Ss'WH0 +&c. we4wզO;^gy>Ϸς8k 3 r 59gl0 }X5g ipxaA1 c"VcG$fD*E 䂋*,Yydcb0Ex)) nȝb΂wD-e#advkjwm6m-NJ@氨NfSuʒIM-tjџ|xfTa0aPA iZE5VHEJ8kɣxfōTO4ϗy{A^c +Ԛ8%q@!%yBlp҄!͓E$n$ z8YsxWaDs?Y<\n۴ iX YJQ%HэI} np j b雦؈tUIh)U)x7a1"t>VY1V5 QkT ϻjrUCXjEJbty x9G7N.w5ԪF)?aC"1 cĨ}`{MK`#h&}$~[]y Ea ]u&X7[>U=zo,w#%Z}nzz5?ݛ5'C "-;Zl>?~3&Ɣ+G1\Ξ4%?&EKtΟD"̺Ϗ\]Su-~^*Y_w}n.TONLvM#5`NPُ(m>\;,s.`DPS$鸰i=Ԕ:ޫ**265DUȞO{ZMUGN^KODq >S}8-WNDW>t8"L QmxPTG" _D",Ӂ D1ұ@` {|1Y\兝R_١Lh|7VeG8@Z} Q$E JgPdL- ئwˢ(گt?靍76ި{2ڣJwUjU}'Έ /^W[R4؟4ܙ~; tq^3<9Ld X܊Y +)+AJO0$B\-iBBHK[*ZB`;VX*m]`lx+eTB GU,"[ AHRJ2LwZ ZFLAhF򑖈69VhT0T !F⤍T1oRC9*HtTDSV\B@[7>p_an NSVYU9wj:߄Tg .*qzz~n| g%XyVhKiS5EX$ =(ǚ'Q25RM &EU"]5T=z[AwBʡ^O8,N@8Az$%% GB BYoh V^y{8&G|i.r\u6FvGݝT|F8Dn춵$3O>ޔMr;{Tx`s6 ubɎ0+ճw=-"VMw&5=t \L-]*]ԥ3JG;Ƕ]}Tp ˚[%C]}ކ{<(L!M*n|x* \vYp0nXߍ1w7y"~ٵva0yjK "32x5rkl֌J6]3NՅ,.d.vc{KkX:%CHJ({#Ab4٬2K$#uK aLIqRRڀF\'~u`D"B4F8j,/NjOrqF.핯+*U.b/`ҳd](t2" !`=CFVx JyR}Ž]QǮv,w,p"' U9^G?>POяF9~\Ę7El7eS),M)Hڿhu[#Uz8?͛oVLl5' |2=qן%0'X(iTXęV/7J&D(6?<9꾂G[n-? C jS*%eMx1z\ }kyw켙^}x.u;~Y1bvb^RyqN^Z:}r~~a3tAY&'L"f=WJTN7L >E̓ :,I899tu)M2y__kZ\&HiSMBդe-}BDtds<礰wUUe]Ue ;dWbbĘJ-o3*bΑ[Q"8ѕkI9DVn%*ұӍ}re `J9BM o]-Z/J֫! CŮ7>:[-m4ݛ\lXqߕ Gz$l+О]S\SꡜVB)\I,fcNր̃ /Za%1ɔ>L3}̤o9R4 ӁmۮD˽,t čQ"i,dcA"M 8"!1BB0R\x/qzfې?0 ' Za|&ZPrj*p92]0ԓ(Tm)(͢,(Go[)2hB)dY'Х5lxBOI {FꖑvJ"; %sc 1 DguF% "z0[BR.kvc@6o\Sךʟq47+ Ӭ_4$'I]^J߼rQa9=]2`\.Yiy7uj;cϴlCC-cq` ?=2dW'=k2tV>TM>l0___T}?~~?z ܾ}_/޾z#izH1nYVKN X_Ϧۋ_n=QivjaS[cy^'TeهTrƠ^mH7/~ſm}lL]r1*U(V -t5Va1ZŪӗo!FXΪAb"/fHkGd8$K)`*`JA樴3.`ނz;ȫ.=/kA`:kш<㮤m$V&"{Cyu:٫vNi,;4x;YM;k7]x6>tqBwOt509v 6 Z URv#+cx3Nk&x\DPFBQed>J\hZ0YbpmST4` 4wJ.p<2`сCf,J$`!(͕ E҅nK*o-gmM cb:t2 bm6(Z<󮫱w/e>`?E ޫYtZZq6,$c=qy vYY&iW#ig Q@da_4XTȔAB7.BBF6q#S{$&}}V ^7Ɠپ8:#}s}%KsGr0:w̵/%xDx.k7ŽӬ7'eVl];scRBٹBZj 7zEw*[$LhsbHNw.jfA26S qEm@ a6'u從7n_ aldOp\t@.NҨi^TipXUWMuwh (k56ݗE_Yp8v-kA^_#!-q^ xxW[9HuVRhQ[#M-7$yR=V(Pg̕SZP3X,(ţ,VF80!ȖIQYPǐ 3,X`Fu cKq8/,ǔe`מ̣5[)rr:{ )Rq"h9iO2n|Κ1FR%dDt(f Z(JVIUg(IIGUQtZ/g8]5-9i:/CGtY(4?PxX/qPVrpPi)Tmx8{U]˰>Gh~)m/\\{O5jxי>Wҗ]~YYoU9_˔({1.sM1Y(rpod~jUh,n'T"tst"شduwUwjyq٘V%Y h<'t|X5_a|0b~*/?:J1o}*ܛ 0>> ξR7f1hr3UfŹ/'B D!7CV$lns3o%նӕ; =C,SwRH z^/Р?ɬjzr>k{%NŅl*CcŊ+]9*|lg#<' YzսmECp|AD)lef G4t6{gJS[¬5G2SXP \KS֫بtQ AF)bR+MU\F[n"1Lr1 MgYtH >Jr㜴6!S$bMF̎YE0 !W3_!xc0@ E'2:/C&9șHV oH+ac{c{}ޝ"zL ܼ1o=֘&oԽ!h@o]:. $eZ[KTkU} RGH;mvvk?zuVI)8<;|" pE C%KD:wXwo@w"i%1>m2M9])هn3آd1U98th~T!ugCHRzv? ; Z)Z 2+e`w~^{OU;vV&lԑ2r} F%3:HDǹY7NSxo}y9]tc|4xٻAٝ?Zm;uUYmSFf~WV$&s Fa&J%UN9&t^K-h&0m2ٔi)8CM , *3Bogއԯ.c \٢7 HtQrޗ_UI1{E"31(NoxX9&: #UhZrFU[eD]dtIcKNumAMbJN4ޟ!9UQM Ow8x^3rӗE'gMi>=m?ylV+3Fn<xmy;g>w`;ܬ3Naȿ:< 7 c@O^?yda:vbm#xL !*b8"%A1$'ea 9! ŕFAK, 6R1ɟGL!oB҅Nׅ F3TO 5{h}`Tw2~4 [t;YWpp-? [iwl؈;R !1>NS=3瀳LLr) B($S1f4jϒqbĜEFXp4gSy^jNּ'=@&e} jQ;| iAe޳B邞M۳gr9H `Φ%VC),r-=VOEc溤O$),yBIaɓH F㣣"*FS[rV8IWlЖ `q Ѡ:QY; :#{g[uH ÌМ-FJnMFla<+Q^oa~>&NN}sCGvֿ".uF4 `J0b:8xUT2D<91j .$ <K)`X{|%RS6cƓuqC(AhsH 8D.䋯mpDRT`4TDLTi+Y9em C\B<\:0Q:ڀ0ˆSʵNp!C~b:BB4*kc`2 M"ۨQ1,N+ rGKpk8x8ƦwV:[β|$˒j߲Iy}a?%F0*'sșc&72\SOc/:뱳;Zhq 8< Mb 딕ځe9fQ&I19% BTTCϫ76O`2 #AL.0CUy9؁MSGZ_O{jhaQ(pRRGmq6Zi0Nב`V;r`*kq߳saI(viV̺ޟu k{>?` k|p` E^SeqgKi9c5ip9SHi{+6CA10&Bp^qaU >m{DH`Jy\pbTWm6k梣iIo&G{eIw@TA(y)RlՌJ0>,&̻*$MKH H6g ?io hdixQ#kg/J"g4PIw/\%SwWM١ӱg,ڔj:6ϪgeIڇȟyveF)?a-Ɯ.xǬL`UYF$|e4hM] .>fMEKRn舕(uĵ:JqBP9֋Rj)Ջ,)Y6I'x'`GDMUjJ)M>aC"1 P"[~0Uؼο$hJqϳkN'7D .φN d?_ӡgU9GW5?I_={} *{~N_P="Ҷ35qx/ecc:t_e4]Iџ6;d~.NE/歁p G> W~zwf.&{~ _?/kv%"][,tSAEdmm 6KAVSNڳER%d]UDR-t% ܂: n673MOje-n2Z &;\ns#`DPIr^n{[&jgUxvtm~[d5)HlnPm0W,3X0m:ȡjg !U BTEDst``t,P 668ъzrˋips{-eG8@Z} Q$E JgPdL- ئdrkE%t"Hטxcɡt(/_Zt5*wqU򊡔.$O#?}dA5??TKPIXQx >@y|S(Xנ\]$y1߮;;\ ǷG,w A ?O7p<8jKWQ[_=JwtZŶ}`6t][zUߋK'ܓWՅUxP\U (@ ` MnЪ VNlZ^>D,T(o-sY 6AlޜO`g#"g0rV֋RIW÷$&#Hvii?Pe Ri2+eTB GUTu x$ ) z%o;cƌLZFLAhF򑖈 i!J+Ϝyҭ(6'myʙPAXNE8e-ƺKsk G߮ 151`In'6(nǗT N8=] 6DL\s<˵4ǩ"dC-tEq47NǬ2j6Ť(&=}2IbHvMH`lp¥g5% tX@qMF ?HΉ ́F+uQ* WW\`y^O>;-kK{p1ؖbNV֭;Kɩ5քP 8%S(=[Z7S,ߣe@q]sK6d=^)BBVʀD]>2 q.&: -8k;vmӻqVu -CJ_vwXJ{<(L$U;:rϣ9laE=Yn.DR;0 iD]`--CW^y}I`q|G/XyP|RF7RQ$ z.BX!4(at ~>ؿh8JA UyɥlB C ܍8`􆯓=(T)UvJ&oޒ|KBFo$+G`z6=k~- `yl@Y>';`Zi2(ø}鵺l?<5/Ljb#};`SQ9C)*i=XXETY`ᰡ ^)UVP-x[AsP[=si:+]{2#H4!q^@7j)RQ)\İZqA+o&XUHy@,G)2F8XLc%¸CkUdcN1~20"ۤs*5ewed7J7nx)FΝ 9IIHƜZ@R/83,c$D`f;anu$Fo1s((5 Hώ aGc`f]\2jKc c6xTf8P&;1[a@ }).BTqj|JF4L B:Ikp!(by ;f=r\#T pTa8F'e-p-B4P[*DJǤ Bf# 8I 8CgFjv:>'X6a87QDj&HXZ  D&S&_2Zb"F5 SVg:GC^!#("4P%p4JF &%WAVZ)Rv LvA&1 >ޑZ^۳~Vd[zlNVnǡ(YjXJKIx}V}ƕ]b-@P:ZTa-uw=m'-M1UJZ ˴  l3hBE+Zojzd|ҩx*drʴ*|̓cb2i$/ `W #;6ɢQՀ|͉YV cV#ʞX,`&Ď\"1h]1YWr\@gF{]` bXKY Ⱦ2Qhgd ϠdDC@(M VWpjᐼ 0|I[HPS@H"bW6;'d2Dj8>)]++]6V;y{*ȋEK*z<-ex P, .c\W$ 8UQ٫ૈ;#Vyn[&KZ>NOMp<kv_Iy\E4Χ,},)VSae,{Хt13 EWnB%\ \Viz5k)a HcJɮU$\`@;QB0v,p b=zMv%NgzLWh" K&K N+]Kڀ^uzPN"<)l+FpXT$fJRd`?A.j؃X0"m*p(X00l" wUϒMDt6@k.ǟ?HugH,25jR,G޲PR]n[y/GmX@ˠQ HX L%/Vp7$J/ RMiX@zyk~rr>ۜ9[-.'͹I4* nfE+= 58)tah ENeoOm*5fk|j9GբFhnƽD)O({׿ eWZTHs8nJxKTl SbH|(GЍskPX 2*PҠ2j@rHIO!H4rboc2l8l 'g (+5҅PUU;I4exyk^+J1-FNU(&#!,LW=gU{l! {6"\4cS$g6&g2"7 =xҀcuz@JQZ^fd `2 B$jd} Rtrzd-IsP|P<f1Hڊ݉0ڀi#b.t RUtiYS0qP/ZCׅj4>)2`p rvXbčMK}``eSFuL!͒RIB8@wQ-0 UeySGռB;]u)XoV/â\M''gWX#ir`gڹ8mKӃTxlդW~v"Y!f O0KDH%fכ7W{s6oc=k~g*܁ϙ;0Gs W̝o̝[̝Xz˭Y~Q<Zf$R1G|{G\&7G/GӖG񀚹U5٘u ߴ`n@$CV!J=Q<HtLxߙkZ>.94(i*Q5^Pve}{~\\4v/Ro1PssWLf`h n~r߃U烬վ EyV[Rgec2ʇ,›HN1gMcի9P"{ssfUnq\;B=r჋mOf:;gdlƺ7R!J͛A:.T>[G( wIz_pa7sn·+Rbĸ/7#ю82Y,QLJ%[J̄487%=oӏOڣ9cFfW3j7/yэ8NR'>hH:*DFID;)@Jྷhȋ&C9SRXGG Uxk8lvB j׳y*;v?rmV StngB-Tlcqב߮ߩ˃>V1>;M1s/ ]zA{`MtbnkP%~veī|6߼;ʼ GP*wK{q >%6JPKব҂Jd< Yjʶ/=Ly=7e20_FiuRNCiH9 )!4ӐrRNCiH9 )!4ӐrRNCiH9 )!4ӐrRNCiH9 )!4ӐrRNCiH9 )!4ӐrRNCiH9 )!4Y)''"I9`H9<)'V}rb+y )XN htN 4:F'@htN 4:F'@htN 4:F'@htN 4:F'@htN 4:F'@htN 4:F'з U:>N գN}'~t} @zweM?\xڃ?WTkQ,(f@P(( 8ՆL>{*=57{U==m֟DOMNmk/ڕKU(ވL/ʓ]$9['r(0,P|3{lGV sKm6Fw~xz)ҳʽD?dg()%*_oŇ\?ҬytkEg{ɍ0ΘInµt) J̹[6mϫ)tCI!΄5W}Uu6S/q)L^lREO.5$ ‚}FvAJR /VS__Q~t@Ɗ&m(%ŪEUڜDRk>v(s%"j%xE[l4x#uR ~jeYԶhHzjJJ]n[w'>UƬ~0Oe8I[? LX5?yGe=͠zqp{ xUNޮiO+ r}rvB1--mcC }QBK(^|<>XW4ן|< ή1FnhGWJr\L?Ny:.{TBL;isnom:OR'QJ~8z=]J!:^]yeu>99|%rW;'PS~V(^'~^_lX|o^;~z)O_3T?`jFa80lY\w7[o}kg|[~[7;/2=y⾟J+B;ۺ ?{(/ ~zHˋ×sm\>h+0a~^/%VVcBo!U Sì>:纛cpI1vrĝx~@ yRZ^8cZ I[(c Y H$/Mrr{afm.]^,~y}AJqߺo7;~gsb~fxu%fc>:*i.e*py%R;sx!X鲮go>}W߻5t-g;⿧ `\>gugo,Z_'攰]3jPn@2q-^CA_NCMa,/v >&j,i,Zt"3,dćMv}>O&gVh0.݄^(P$9"|ol >c_7ٷd:3tܮ11;M}Ĝ1rk u5ԵPZC]kk u5ԵPZC]kk u5ԵPZC]ٻ6W F2T߇‹ $ޅ &ЧD"l[=KGȑEEi]]U=VuZO1^N3U"kn7qكKO2:-Ѣ;0)րH,=kVYF_VbYY6rXa1;>)\N~ҍ4a36sZ﵀[~֙- lwºp5T.8.P̕|Rl(* k'_\^C_$W *i?7`Sߗ1޼햀@ko:;/24<\T(8ǣvjPqn:f.J^ac-> __Agk][zP6cs-8a(c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3:c3`3Dn BM `.? -VĶldk|qě/Ct Nr_$FR)cRc^*[-5ER8[G՗fCY3$wCO$8``P|6bJl +%+PΛo5?n_-m;+bFoZ'X* C!)')@ 46F"x#H^hᔡWNTKe3bbس5 A t!)tIV{6/VR4rӟ(B8}wBvԵ|Rm7dw=~>WL5 [+^Cadޫ˞!|lxRl9 -]Wln5fjnv >I?\t̥s0! \?*ϊ0*Xeu".j넢Q1"} /]a/jx5}}(ըS:8?g0x+*/[ԩ09etV1|%iغGV&â2^dŖȨ F{ڲ#ZްSt YX8ܒBuy©}gvBOgn;pcDQht \tq;>0umuGL|z ~5#XΎ=`pQ {%z0kR_~%)nLm^tvڳkBe"M;PXw_W~N~||C1}1:9W&mbV׵ךdsru\~q||28XXNi 0;N)5Y+gsZ?WDޙA`8`NDڣ էG'd5ʀbǎhR,xZ-^jMyo=FXhnl#E\j$)^%'.)Ny.1+T"%,pjvJ ETE$b Jj )[*smW/זjc/`^n'.cJ"lKh 8F%E vʁwIS#%KHJJ&8`meY@ēh{^Y*u5%b8kw=5̘__Jv/e%\PZpcŜxۍxYxY #tOؒ$/JSU(DU:$d҂mlW`jP sϼ?:Sva V+ ^M˫B%* }c_PiH:"ΐ "Kc"=4u"W(;,k (χJ\c>Zmhep$JyyZaZN5A.8GT-@Ns [8եE[miK*Y0R$hGMEu4fT^1%tEf$쥦X*dUsR"&b G,˗\Ùx6]dMŋ pߴ7 ~7aU$`AvjEc4Z-((#,z;MFS͗o"ד> nd^ \bf5`^Yz}~҂W%7iQ 5(¿V8d^Mƈ<,FDzP؛'r(ts! >(1Z39EV 6Jh֝ #8^`go[@'#GxF@J0>m>tEӇvh3=w{읷bd>Vי~aym'fƫyhr zaz[a/EǸ.GqH;Fb ø5q b.e4dl؋w׫k:J́Qs! ƊD|_>vCkb:Oz9J3SóiBo*zYo'OSm̷qU_NM gQ}rCŋo>/U|P*[w)\_޼<1 ŒWέ.#IXF,Ϋ*xV6I/KJ|Dm|,ӽf+6Ecg-k )݀/tJÃ&R!T\XC".3f1iAe:ѣyi*A7:dScUdaBQ'>,58G!wLW{Acv *]J_ʺq؎~zׯҿ8͇G?z:zG_Zgf`(NCwFX8ށkM畛^Ba޴iM+)dooWSڽ.>SNj%@?_vGu)OAfn2q b~jiMݤ+$'Z\O\~Xh)dRɍ$v%q'?ݣ}~!X 8N*0cc3уQY\h QdVT@Zpye%=_mˉ)gRFR##`cl#6*`{5QK)aI1^:[EڞxDHk{䙛zFt٬X2ް2\I&Qg{MZ \P+ V%sZjG rJ|y AzT!$UHAW^)ǡ`U@rXv Z˽G^ R+g#e` , mL =8 }˖g'9s)ʝGv+ ][oƒ+bI6ݗ,rd7 Hrv/նȒ"[eشe{ ldWWun\viV.32~\JiQD[k]޴&m :p᯵ ^rgey[;œf$h൹}5teZ*㷵}Q_c܎\)OPo n`=?+7l҅{ma~n5[ͯ 6nw 0J%4:Lix-ɲkqBsyJdem Z`JZTh+T<{m=j&b+Ԥ,0+%RNVu95xc0)K)zeģ$: `x2)f)03%nV9E:ϽMJ.ِ@Jr]Bz5 M@8$`;僳v5ȼ26Dn%E-`K5q6)0I&{JT iuO {2f br#h )Hy吂AЙh d3]ރ"^zJ ܽ1mhkL~ Uab4dB\;om T:hP`E=l?QdY>Syg0,SWr0 @sƌйĺG:0;w#]_~?4pf:nFddh:j.zqсБ_;ף`m%5@J,XkrEk\c lNwڏuk3'Z'BS'B偝UhM: 0{)Dm! BV~0q*&[ŎD46$e@xRz+td\ر5v\R6z+K ӊnvSۣpDȷɂ0@c #rgd\3<+n sY,T }) ZyA2BV T=pEB;w u-V۶H,lr1莅n"t.{c.&/Zt?ukt\,w=-gu32l ^D5Ya[t%$x"&ObQ:EԱ2f21v~c3<P=5{aF 4/IU8( 2')*pYrFy0"X9'J6'^~OQՀU;PwtrjhT>γB9Cm5uךTZ)Щmh*Tɢlhrɽt˂EÕ ,x0!9'3bz+z~-.t/WHWy޼|z0g5BJ%$B94RGՠ _z3mB d|JQi脷] xxyi9v>kepwD5{-,r7m㧋F2{&{q :kIFyN–i: joБpW6G$2x=9+][N&^\l8~hRi!`}}`\\"]) B=m$̒'x E uVxNWzs i4쑕x[x6W[+cy(]_X1SAA*LGT:&>.,|~/N{ڪ#(WwL2%F\TpD0h;L!ݜX/R;pG F944hm&LG\$9sY& qsȜ:N3 f)et@z?{T6",kן [MށrJ9SO{D_t Zs hj\$]'n9- <)LϢ`Jb}QlAhY FacLB % - J#zݍTp:l$<$ܢ+qnB]H@%ȬSa2;{ۛWa6=-W6g+ih , ;Qj<}]]וU؉f91^=^5UO:<DZ[1b-f5R5i2t0}08]rW3`IQ;"31= a{ &n><Κ˲*1եg2gycCT_>]aH8c=!HPsR9ˣV:qҦ%:naU6xq])y6 sC|%yHozxع[wW p㬮X\{ևů<y#K|KgpF0*h\`L4S<{{|ޣȁ\+,eLxI#yFIq(ML%$lVHC4+ ລФd{~O>¤O?iZ-bbZ j[4߯㼯r+y-M5dj"L L$SE)T깁Zoqc 侨w[300T1ӡ1w @өhPٲzw[3R*{.Z_/ ׉hX}=.'Dt 3%$KNZ2Qc>`="XD)Y1)A\P24Dr6((ROrg菉5_/wB סE? '=^e3l /ċ/$AIIܝr5.l\e e,eƽgX.+0kc ^cCߧ1%pxet5fo7'GdD@<'aҽyx'`NG,|IEZ1跲 `3E#!,a-9]AY'xNj"Uqɏ~D,UP+GC!>~GYt,eq0"0DrnA 7QfEa>+ UQ XZJd'mDӑ馛ZB)M bo=BfTawߪؿWy*dQ?|;3- 18sHps/,u58]۟_Wh8U)E_n{uϟ/?bI'D1;keϯ~{fs3?)s_5|wxKt7gסּqWb|t}u7_6d#*a}~ߙhhtJ8H<v:IR NyMaIq'w/CAsm ł}tUн#)Mn}U`(yQH׮61K):!mrp|MȆPǏ)N |4rWn{W a+>!|}U^*?bVvc7Zt~}a/ =MUl6xC=:ܱ?uFÆ؎>M 5nM֞](!ݹ'@ȭGtͥ/nK6ߐǶ}usWZ7}m%nG;;vm5r=> }Ww~9OyC;:厜]Ss˦+mJq]?6./>}gAob+q1"mLuyH=їn/X#gȠ){O:bJJ9Dp? = io z!9։,&SREȂs%LʀsiLsp_o0٧mQ|?FRH%NQB&H,(EATxʌNh9)L]EM5rM]H*[I+^~#! 50*+j5qv+j#5 i0t ɡ r>8a _WlUqY [ƔvZ`1x10".YY^9H_BI[Έ*OYrA'21;'VIQs-=+TmXMݖVf ͌Cm!T0£Y^]d!giPz q1|1msZY,X!v nb G%;`[*Wٳ2P=MMЀv̺\ әfUjb|W\SfKh`KIUHe+B`$q墷IZ*i4)J[>+m|@8LȐyA&C&/I)lFcVr+Wg=ybO"VZD]""qkKYG@y*= :48E>cH7Z)|=qK5ίWɃǭE?.僞tVUWDJWs!ZD`I+(_!>ۺryE=79D'1;B^tʣa }ɹ : x%yg}z2ta`O^K!saq_"}&׳zKZKYw>Wo59!y` ̰.q=Y3/6͡goLs'L=uC> UMXLO8 x/GLuME )>P#H"q*esȶT"ć:)q`1'c`d4 w BGcLK.т`Xu6dN(ZbpT`YZGt琝W*%7)$Qfe \I9c<0V>Q(IqW坐y'zE?Ls|r=P+5cgZR~|Z|?u:n2mH^6(o4goɩ8f_+?Aܛu Mm<'mKJdq@2ps<ź-?nr4Wk5NVAkZz<{/ɨ($G]%#5D G!/.} g^G).T>HSd)Cwuΐ5 YK DC!&sm`TM:%t IGCeݧTThi_\B"|| ^pw"8Bo}pT>q֞u/  P Cn 7 'QPg,y"U U 'Q5dIJcr\5KЇ:p b2a# B7,w(+lK )ύ>_x4(c1Bo1鏪GJj{䲏`-Q+LJK~! n)+),L˥`a\y~uci4A 7/^B2Iɢ Q UB`#jx,*(ZOrQg֯bJ^YC :/ %Wڞw $Q_Jx/ <4z]hys3Tь?dtЅI9>{75bO$}*1~ N_~Ki5頤ޞ=z.}75&7l/yE\w쁢M=WE_90%cbn o!^|{'Csz[kSV}/nJZijޱN>wY\ :n!7 ~x@w_ `I?'I7 wc$Sre S_jc,:3.9&w7룛J놃Kn%Pۻ{%sYR6%tpx8ւȖb .G)m Iv$v8Ζpqlb~q_lgӣc9㋃E)mSin}os']N` {C#4]V0@C{G)ΨMOy}oh7S7!j8Ρ1UzlsG]3nmO3Όe5YrLFRBK: h&&@@(x2c,JH=!U@f/ֻȬ12">ed!F2Rx@<Zd̐ns|QYLiˁ3F$?׭cR)ieɮ< Rљ%Y;BWFĩZll9 axPYYZ>ZRɨAL:kZpc)&olY_޼zaLBrJNai]j6=[ wF?+SAoO//|L3Аxq(eU A vۘ$)?b%06 җ92;X ͩ%0 !|U̥*|XC'o;ҧTJ N8޵q$28~` aZ""I_ IDc 1vTUUӶn䐬jsc'7Ȑ 2 fd T4Dl92ڟ!۶AK{QA=!L;=j7[(|R}cօ!V*}D8dӋ֚EE.~=(ٚp\a-6f-ccCߪ^%CF,L]}Y٠_HK]ׇX瑻|w]0n[(s5E 򰛖ȷ7]Xm#?\ƛ_mNQ!og evQ7: n|ޠBf!ۖ0y5S}^.*H/g9PBKk zZ)Cb*!"|2)Q)FXBSqWNWw%= Jf/n:뇂f˶v{LPO0.{zd0^ܴ䂫t 9ϴ:Zjs)k͕\;rD}r1LCG(0BD$g$HP9y^*瀗msacU h OHe*Q*2:@} F9ᔕ@T<NX+F.%d tϽeKϳ;|y39ӢMG:{ex yWmG93UʟpYʢ;Y'ucP3 `\흔5hjȹalc{{0yGheLzhn.aE9 も4 fqeOfFȹu"Dm#*mu+b.WSy3_NOW1$wM8CAri "3`$'a}` JpT&o ,ZG ?L"f8NG+ i Kp1rKZ qNN·^|&<6E8mW3Ycw5hܰ9h&sJ7MF"|""iyv ݹ98Ai ThX`="Rpe4I$0(LC>tk==vC¦$g?y9kjq)U2% I1 Cn$d *-%x&y|ɾe-M7aDŽ1y2^܄}4VY#q !t@JԸ{(j!7{C>մ'}ڞP{RV=iwF<)au~#ٜ:zCѥ9Ϝ&>imjH!(Q"E=DaRk(c$rq6AxgKPH7hI,/FΎ6B̡^lV2%4WrU `vy>l,_>x-+#KtC-K`4UxʅAP;c'$'PبDJA;H.!g[í'_uI>tg۞ƿU L+`ZK蠓6T8 ,O5J8P\f Bx=TE`)𰄕=TjHO 5AzJ;õO@o5m"R4-$$,?~kziA R^߶މx$T@ħd 5n" zA*+H{c<; Is6g%tx^UIN OGc#Isvm8i>ۀ5ǖri) cf`}.Gٽ'0W @mRf|Sɹ Ta~$_g _G'cuZLp{11C%v|NOp&8Û1 m=6(N w`=u@̧DXO.` 7 §/9:9 FL>bu:F,VUzns(*ʾk6^^-=xwv}1Lx %91-jF۳{ ٸ^y\x{9=}n!Fpq˱gz0Jiޮ뒃l<;uaF53 ~iX4tN(P#ƍ̫whx3_휃tY9+#g?d׬jҳNc'>_<uW'=ոa͠ˏrP<[;`^׿~?9O˯?'̧wӇoq9B :HP5 * 7e穵4Ss#Om.Գ>#w{_|p+[WcQ`yd- 9 F (g#a6DC՟Re} ,6* ~te7PI+mX$v}Mwn'YQ"W`NBB'2RTV@NkIOp/b;\rLpyLI0zl>w\YdHy_*%uu&0 -ɤOh|>IVopR/E ⸚ +?fjiģuG;P3~AX!T e.Wg?v <pv;k}gIߎ#β8Ef+?ް;'%r Xg{۸WW7c`,iom4]`78nlɕiヌBU0oG8,|3`z}ZCRfbBBfv9fvϖ]š oNR?vlf*YZmB77NB(R ݡɟmea)ԣ/?:C/1UOFw /5NG@kQBWJoJ-1gŹ.9q`OGNt!ĭ ]zxDG]R;=[JW5ȚJ׸X4Җ*ԈFTel8E\nCt@_{Usaѵ*kF)U&,ծ*RR(""#[X`ktk5ő,2\ª1 Q+IԚ)G5l)J644Ҕo!1Gu]ƔDذ֕p$i X] sA n`[* |RUv%؀=ps/u5hkb$_}1ΜsWUن#42(U0xQZeR0bKm^!/܂sΝ KڈV7>ٲnh]Qġ0YjF1XYcfu; DTHO+a>2ޏ}+ρcϚ4ЭcFG QD+7N?Lqq~j&LRƪbq^y!\TS` b*F. lEȞڏ_qQC,pDm{ٰJjQlc 2 CRT%3]6ZaKyUU츳y8ۨ٠'oO~zRۭxwf̬f`4 ðDfq[jeL1xN"dX q"pXHn0J̙34aS-/ EB]ЮaJv:5|a}'pw3.I/ݯ\Wm68nɶ6mWh=JvX%*`}9ͤ.x]zF/m; 'og8[AS<zp0>ힼ\ƺ:Qj|1^0zŨd.ZgK/~![MN<{*'Myt]`7p V!]EC8cI6b"sy\_+&qUUK FQ]hYQ)E 07SQ+6X1eݭ^MʱӐ9+T8bw,meGP,fVieRVAoJJJucjE0@ aR WN-BƯ?LGAGRo JJ:r<Q*Y*.Ͳ"HeXq2vlR$T-(&JF QC̆Uz ʇO:P ;xؘ5-R:pyyͥY$Vt^bD 1k58c41k쩹8?:5 ;>-fz::cĀ?9xj֣ʛxTӝ{e,ص;|[=`pp-crk2ݴK[NAK~da潆K_]xKCPBOO;.5XwTFuxcl-7X=ζ>:G\M> &RbRUO@]:{KޱwH !lHb -9٧ )N%I5ޥ\X+/CP.zi52dYaBkZWEU i++ӸO%Vl"+5_E9]øpAzo0&_a`&@rXxYN.Fq-He׻ ظtR}L+m09-.ϋ?ՙ"P|1brujr2W3")S&rH- [so_/Cҏ iEM^pHƮԌ;U ]>qE6!ehUlm*l륉tzkna˒@`*ֲH!x(uӸO DREE):V*-"o.R[(3uIkZǫ<^ak&2_G;5r6 Fz|1[CH J}v9dJbS9N΃:q'PÊZk*2pMZRL$>Ǜ}n$Rp7kݟӽp.Kk|6+xϨԍWs{޶UYI^M˓e WxÿzD[ _q_҂eSb~6;u]%]]f/FFT+П][,XeaḴP\idh#Jm"m۽}3:+JߌZ~3VL?xx7z63wWLq+%6͈9ȝ E;٭mNmenQXQ\Z,L J@7j9W<~Dӂ`_~gmj%-ۻA[8vAWT.L2,:"W!ĉ,9C(1%c`q˲Ħxs4c_Cݐj%"x6f Z>R f(ao1tQ,[b{yV& Q\"_%ţ8H$IbIf Vl)(o3ƬDD72}2L8$ҩ@(tXk$y% cL;n0TddSC71u`iKρq >M=.<;!i^s nQ+I:NN1Kq>M029/|U})YO=9lY]B*xb̻c:z3FL<*<ʎIY1LX$p.s,1F&?}$0tKN#[ ʐАƦ6J :L1ZD! taS1e$_2DqL,SI`  Fz lƬ1ʣ "R*Dr%@"NZe#FT4Pu؁1$#4tə(R4O k8 \ V$]i6m.+;7- Z'&\<]$:OKHgmGYQLUCOt?0es?\ Ҽotz>&~w29;7#<3c3wfP;j*'QxIR40̰亩rr]#pB{1-,ʀF߼WTӑ% ڝjJBZLWR鲲QKQrElBa%MYSmYaqx6d _o2Na^qWWIdT>[W8? '_>[h2O`ߛ̘G]jUՕaW,RɲՠTM e%`+mjéhD5i{yVFd?x-$gi[[|ϭm?D囐5vj[]Ɇ,k &һH1F]sEBö;bdIra2A948.gtgw@hap` / w@|lO#ةϹ E*Q]J!2)C#iW;!Cy1W8KIJ)iw  o)* S]FĠkXDRޕ5#LDIEFLtǸ=uԁiIITRF%mB& d±X z>YњUJҞhsVF#LͤډY)+2&e%=Y b\ſ<}´?tc%iɌf@Zjz : Qy1j#2:uswVB~D*ʂ+|z߱J@Ce (H4fL[.mUCUxkx-2t^ۈYk%ڳQK{RI\PjIz+>mfK."Z&5q a͍˅mn>F6S2;AYvp'\~t[u`ӏp|q7eڪAD˝^\̄CK=9lw!s[΂G[*!e[*pv>]uQ+sgeU, 1CI[.e_ ͵ד,bnIf߿Vo(?ko{:R :=R͛B[*6fufTڀ¬QR_Z,z5٬Τ5`d ;'0[ gjIsƐ˱+Ĭ5%jzҘձ^B$}|ⵘZaJ}FI`*1̣q|g gu/vx'Ag|_v˦{YޞlʥE:y+oY(MFY T) A\sE? >Jͺ˭YyivFFZ̿;HB Јch)PUi~14S PO",@,>*$W)B  NT1wl TkagҳLU>n;+=B%,P.}bEΟUUm`2?:d@?]tޗ֒ēw;;#VUOOe&O3|u: \pv&{80}Lt!lh;l:t7731˗N=$ƚr`{ףⱣ?5YiQi$O7fhZ"AmPE6fǝ]=/b' H@7-ȳ5)#{VrElqtXş:ژK1Zy6ƕ Q@$Q Ix܋b0'$ 'Mv鲩mcU8Ȋh\i;?kII0z$)MIw~QeRڒ e昈#;+sͬGgj̲&MKՑCqi4mt×8‰c6&XVk6MT]D<$y>}Z 7fUX?:j]a: xU?͊1N/|E;0e[_I`kGPz8Cug8Sq.rh,S cBse1x1!2D4"घIbRg ^Rk>11OR!єa3 e$7DB$ڠo>5-u`rIO-1<U`i%d`;H#JXYPCjr*3ceKF<^aaH"1~EsAc*bTg`e)1B$H. < yNb x[ JL/K`tya[38W' XSaD (OsIHc:羗,c 2R&;gbISLAhDsQe\s\M-N/Zo~jIY[q"g:hg@X) g(ah!$lj6"8``< !jDJYbK ܀lSNrqi1JX#فٵa(ɣ1nk*ϼ.Gte2A"iR\M Z g9PL3]O٧2># +ʦI(t^SXn7]{anݿ'E$jPmU :߽ۗTf^ZhZT&G-G<8uĆjƩЅyMi+_-"H *_/>Cgx`gOFSp}3}Vm_Due(#㾗]`_H RHӖ®/#9m:LcY͡*{!p?6ԄlnQfa$-N>W\uʭigWl -Ld8>|/o+^,L=[sbJű/*ص`^8lODZ}t?+E c{!P1b S'.@x9Qx** hghm1>z?W>w|ʝM^^/x"9:|$)~r=l -oֶa璲w|q4C׷ nhx:.@0:H񞵶(;˅l[zX^ JԑfAe3Rrsh$"Y52e8}H Ngkh1tm\L) -(7{Dx4 涂0;- d v@ V`Y} p.7%Ͻx8YT-*U)VrZ*.ͼEd/ly;st(MLaݖg<\f\ ïz)C!)u( ,rÒLCW>Y^ń$>\@L|,>WHlږT$HD.TDS1{ E豰ds(A(BÜpQ'H T$&K4IR+kh@(8jL?a@))M,wY#q)%v80pV*$Tjq=7e (] 3P#- Bzk'*ƴOP~఑4TZw!zȬ84BargwGj,ȱ&M=t KELX^6 Z4E@KDvSSF^6ZUJ%BJS5CZK|]n"%u.R>|lPe0|xҢDЮ4׾"t&a|#FWPTIyvlJ=Tq\',UIfKjR71OMsii) -^!u̇Qv lwg\>޲ԋ+`ƖHriʊfkM:IT74YOȖ#8\4D.B"|{T_<:і5`}蕒4^-שVy ߻O7 5e RE]hAn8c⛈&fq#X[hXsF%_ֹT({\˜OEPN+S#{g(Y@% KtdX'ת}%r!\pz(B=vՖ1^aƸ;wJ|lxPDVi%v/ecYo0E[þӴymG B/'sV^'2+1EFɪvGllxL/gdSr?dRqNhq~D}l8$X{+YjG dGnq[{I5lEw'x܎< JC7!ڦ:Nnq %-8y\{Q<2v 9h X>\.pZN&(TtE4C@@ԜGIDe[gm55ZbI |TQ4mTC8EgJ.&TPH?yac=NJ0:J_G)<%HQ"^pY:jPރT79DO$Й~d*xy0;.j>NTv?8AsתըEaJSmzꐌ f}L ♬|*/g .4R!?mPMz3yU׋^qx60EY:{.$%Im؁'3Ù=HuTc1МWcvlA~d|>?$jq+1`mʺ]s |&(B"MaD4(_vofSѼ[g@i1Ok>єQ㈥M,KLSD)Aƌ/qoLU,4>!z1BE|g? s!ӵO@( =?Ͽ_74zME @$AAm8K5%s;HB-] n0I 4_gݏCI>l>:< ]T7LcFV~.Yw?\J @bk|<1+Ԅ R$*C3rƅY& OY ..R-YdmOqdq`$,6;v"n D9d6墚nie ĉPHD"Қ|Bu%ҮFzBIx΍@AiZp3 <_NKL45&t0x[wcfZ%zXc) Jb KU2s(~,1uT'΄NyRTnjg!b,G( GU=Ey>55J4R" 2QPJ@yqD O:u-\ok%&Uf%$E*)0!0%.HI#d*KfmbeMu :{wc#zGX*[lm_X0V%,p~ej">:ToO/}ct) &#e^@ Fq ͫF`J%0JwxQ61hV>`Źt5ʼΫƜ0`Y,< I)DR8wsRfJ=RVJ氒YZ m|~)g1pq@/2/۷tkw(=ѷml8ݓa݅{];IP07/rI$qFdF$9:y'-H]r 6ˏ]IV UʹƨvUbuP_6%'%ʸDܩL,S7C XV`yA{:`W`9U?Y%9jk; 9݁uJؘ2 )c:W$(<ӭ))@(4q;$`?H+$T ,QR0$=+$/DEOpt!u"$p%{wl?f-. cC6+h_T~>vhP$4%$ɑy B(4M(Wem+Դ6Tq!:",0Ԋ{ ʪX=u9տsj{wNyKH|Fߑ*| zuK(<¼'>DfD֘C[! mZlRd3nDŗG9ةD>g8:fe͊*ɞcS(]J=җ# n&޲=f"] #_}fŮn,7_]/_ÆҤÇ"(.t__q&p%VP~j4a(V(ߺ-Ѝ5ȥD56*){Wv9egԲgΆ[GMbp97 OlrpʒZ&cÇOeS׵F})&M"˳$J/ש9 %ɐҲߤ&NB֢$Vh}"|N^& JXyKΠī[`"ݶLNPJ5-%o,I+)"پS ^ jWtF`RgdL & yl{.z1Ro?թ@:x*Eۻ{΅cc vӚ1DV2/u6Nn wN?y֨99ggSg|e4"Cs-Ivo X\p1($Um).Bw56/vg& atκu=ĩH&#)@i$ݻka~C<,gN/m(Cd*IiWӮWcc ;`RM qap--0JTBƽy7h%{A81י,2A[D rN56Xmlh^l׏٦AQӁxҡ WRtr:kl 5q֡ZnoYQcc H0\a ZZ. ;w~N:5\43nJDpIuSr3a^ 8Lw7!U hCZZ=XdJfLfq/g0Nd"7S{*6+75U껼| q„`ͻaqD1w3i1qP}I2Hʦ@?uE{s,,6)?v+LAtƵo>gZ(ޓMt\cEhdl#Qw ­c Nϴ[߿ SA*kڝ ,gs+rsVRhnDއt& ˄,*[ e6-jc2Ɂk;>7)K<{@kI^.D^f/HgJ|႟[3^SZXUe9 WeXR-")w|̜ku/b{>#f :ܝ`31T<65iJoK4>ګ\Eζ>fBpT >eO8ePabUh}\뤺]l|-d.wO[0+qOGp߿7>Sڢ AU(٪`ut|n^b L Q*hfCRET Tg k!`| ˀ) []0*88Z еFHlS5mr IVDc92Q]9keiA誩Z<  Ev*XS>Yu;p~(јGm"0wH)(pc T MD`J%0_[Kasz:Xh:(?$ym1,qK|&w8 Nٔ SQH3^Ij&j( Jp]dIm9R#Tޙ[ 1nZf+E^b~`,gAzP)銾W@[&Qu*$ZjZzL:q:$JȀF '2z c:L2mVBj&Ywt[iI8sϰ2)X$/©M3c#'" 3hf ے̙쎺-I2Fr sF"ZaNrg CH.$gNCNp&WkR\ocU齸ds>O*TKETWԒJDD!F 8_K)zޥ/١|14-tÄ| 'klYhU=gjl2X%/' pMD6 R]JVzU\]xl/V[/53"D.D Fpb&os7)7TKlbý$"䦧oD0k BJ6s/.%flt|@s  =ك8sIp>^?GYۘ^ggVooqNVeZe"lCK5(|˰ߗWvL^g>C@N{#ҙ,&_eb忋~G(!S([g"ϣ,Y_[۫ȌdmD3Jo*g~!)۳aA>C7czMNw.D^f/Hg3@AjLcoWR0)z {iH3p쇕/6wkp}ceVIې<3'\p>YRD)o߼^`F~.SAȽa viDtw:3AKt&RZqR2眗)OfkZqizRzàZ!VgUDmBJ, /Bs8|:5oya3_Tlk2ƧLfKvBLr$oyBO`5,dQ;/QY)' _cFv4F@7r)ʠ{``HWjn ̏gеugeJ7ψQ $B07߷Mk-P6ךּ0?XU<° 3s>S`26{@DXq;tgI{DZ+F,>ݵ;%J+Qi'2E&ҕRD'ueғb Z3L'pS$VQpʖ*Z*ˮe=Q4ܱfFh+IcV'(>D@z0v8?h hiB^(ήB f^nnK0Z9:>BPE@~VGv2uke׊j9jr.(B[mN{-j0ۑTuuUD|Ypu;ޑwFedD:M5B.ya?4@(V5H44#LfB|cZ'@^nG>UW*s:G0,#jnCVNYD ?^OnbMAgǰ:~pˇ.nzx&}m\COlM5V?A6tSer`?l{6Fwl>nj ;LHhk J~]tFH6qy?l O%D> E Ʃ/L˒qU_r<^1ۇr4;X![zK8r@|> BMg׿-iůP,r19rkĥH3NSiņTof&(7 TVCJ$T(% ß7z'J%7ha(=̷_;?XKdy4އ[ml.9El4< >{.Cl>`.yu[u=vGuk2f鶓9v{4$Rb)KXloۅF&hcT i4t 6"! 9|؎w)h9Hb^QF },qc1%6u9+;^$pMWzM= ПSM V =iߦN"eր?,?S$>~ x"?>/n븗;f>O~`H㚀?l"$R^݆#)1v8`̦i)xr^n'G+usœ8o SO<:#miCRD2`N]y926Cr'&èQlW<_ $qisTܧD~7x pq LIq&$΅ <ctLgXrj`Q2W*QJhy4fo|N~=Ü0gw-'0c>ώljǽ$ƥ-LӲo|ςSԻg_ˁb78:lP\d%$Y rsaZX21fz[)E"3\!GUw/{cB@#p7d_-k!v< !9UY,'u9Vbߙo2N$\9TN~.wqFK<;D_W)\^(ǼL=zݥ9bsY}׽ ~< !kT6uFLt URʠ`}S%iZm@;lv ;_j_&>ɢOeS,TMUU;WʉiLS8Ήu2&%k!mB6yf HU u;o (H?5O_-K!M c'/fr94c ТU~뇿r! ~~_/燋K wƀ%{K1s[ 4 76'qs#,ݎ򨶭yI0t`Ӣx3Ԩ $:T~Y._Ngu5xޣv˓pmiYhYs-w /0FZޢ6e· ŌIV\+Sl͸m:_TQױ[~P5a]ŝ T'(/ T9|96@bJJc8S+޵$ϸ.S΢/^B`Jbg;FΖ7[oub80`8ъg};7ExR~M50_Օ^@04{Qu)Xo -[bdrH*+) {&,`h0,Ⱥ`:˺TÚBIRc}]%$7@4[Z!kZŃYD~^LpK&]nJ>pJȝ3PB3Uy@|1wmm$WzI[uì=`=<,ʔDJCҖ=ER}`H4U|UTZprOw(]$*XL#Kmqi<5إ4qu|Q9v~Bҧë9{x_|Ymj|QI8Ln ?]ҽ.ywBIB9 ŦYS$lzF(K^0A7B)@wkj*`d+VLrJԗ/v-"נmӠ7\C! .Ko`!^!-O5NB)6GG(lv9eEs9Z3 bEC{!ΏG'4́@> zct3돖*yűUvkfYC`׈N&=S'XufI$C2%b^C@Ovb4AV_t֦FS*/cDO>)+te91-TA. S"ѨCcWkŘRUQΠS}^T`^ln{鎬cRW ;dY ukܪK&,Cä N䬧KS]kUi&Wr/O'iT,a~iSPJ&UV㺯鄹Ͼ4\׎x\ؙ~2UA, otgP_8B w蛸\TE$ɌTHJCǑr!|d ΪC_hq +9X7F{TG \{<+,0&j:/w"dsP]!~cJɡ2R%'ygYh+$Jqv^OAZ U^pyk3Cy|Jɥfgx^5_q iF:dawE?O'/ՙ G ·%XVLBc$NZŴ.Nϼ6MܦHN nHxb n7?3ps7}a9\(S?,JEa,uϹVjF(k㐖T+Par7ތJViiߞ/1&9]heU5Xiq1œ·/!<#SuB+E6w9igHPNbivƞǫq\9L9.;:V ~Pc1nzQbBDt)9L cЪ$`t1}t_1e0r>rnPRm) C屘Nң!i/ZBaHs@BZTPY dŠ\#V>/o-\ 9 ֣1ma46 Ƀ~!7xTZSDcPe8 pn)|,Cie48b' a&",2XG!h$- ?bMU4/BH ŝO *'cz-nMiCpi2MA96 .sigao+Hۚ?8$XIIkC6h0%3NR5_) Dt㔕H{~H:R"7 Bӥ{J#j=m_.iWat3Eas{v??{ׅ .Ma-H=4`xxP=f$uR&.zF=183(ՁM!&Uas70} h-o{ݐ֨xkavx ڡDڃ09'k^+!0O_ݑ/BYY=4^VQ#6齭0)Ѿ1ɩ_u%h[-2+#bwX!8t6b^䀺Pe-'\FJ_r+(Yy{h>_V)au]nnioqUm"mCxEEhJT2s)Tu; Lͪ=4^i7BBF/JD?ڵR{f`̴"-Mn޹|o0%%fEw_s_SFJLn. vXTjU8e#Ah;h>V|Z=pqiyaLɘujJ{Ht m4v-/܇Aq-wz,OV.j{sCt}z]-WI26 Dz>&Pn uۋyH>Oi2O3V.2g追6.~.EhLxo?}:Gaz *Jǰv4 ? &jԔPR2@3 M<+qvM 39\΂ Wj MN%iCLi/6]ٜ v#h|ݣЍ{.VUo&Zg_oÅ my"J*@hyKv0 a:=CM_:n!S-pqWW(ԉ=CӶC$77ڿMԹ ˔SNVDb$eNKln) ??oGpqʌx);û3#G0 Lb(ƮRsLҟhD a. 1k1(3Db@8B"0m;Jy6 _Nj7>CuX# ҟ\Ū>~>0!|uf:>jj iAI_\~w<o~cgǵYGF*A[>ŴB,|m{K'V?ၱп!sx03 i1_xohezïHo~,_UWcDɂ]-f"K!7P޻1fvF`md5\XXN 3n6o>o- K5)%VDžY|b3lJ? KjLk$x$z7^|}mǟ\iiP[~{ T}BZ+ِQ2bh*QFYxϐ1(Qp"ZR"Z96L +Fr#"Pk"p#8Bn3tci A'K+'1xvز:ה7^2y\NfB,''``wV<*$քB`b09J' -*y_x8(/ 4MT2 tG yO6웟}B?Tf}yO[9âT-z*NKD(PAWp ez$gr4.F2H.RŶ>(esU;;(Rb~9^ngP U}cV=>ak!^zcbF2# h! Z"F"閮C\] s!dIT5'Cٓ 8 {1qCH sZ("Xڰ>a0}Vk'AJ#|$*4ujpI RL# ΄EjJPG6n2Y+"[L%|aE3ۯ5^b-W:Tq5s)Q-,* ˟ᇟ~<͗=%?;rD(XW_n~Rb :9e$FC&]T虓3pg<(2`"!Yx\an4=th!ǯqmys]"5a_^pZ9R[HQs*Ki~}[ .\ZjUozC!Xkۍ#G"ex iO=A.e^mْ,[v,KRRfɲ]* D# *o )U5ER`\lXAؽ(:!ܴnŢ81\od ׺#ejɓg?!\Ai999g|7iLQ`4xr֣C FFRŞD%-{ZTQ\H6Y?$_0%jIVx"~C68}Q-"6`MGuL^)BzC^B |эkE,(ڰpR]-Jc-G grz}Р};^`HɧeJ313 qD-{'TW3j\hH>PW+&c8h0l(e^dE0$ y|‘^vSs(_ Fx$5Z.hr,~["ZvJK-m8|_ar,NMM K;XQnL.Z}]ѧeɝ]˻~! B[_x>e må8= |%L $lWG/7N%|s@VdOn #HRK##&h6䆙0>?މ`*Pط%b3S5l" 'D&"Q!x3#M9P+#k†X i M ާdG0N5އU(83'Wc nzmGvp^&&F&UjgWEM:~«^;YdZzYo>go[l .CdnBR+.Aoqbbٺ0p،lua1>1'DLi?FD@ Fks!{Ⱦe|s._ lkWrY~iK#Y]qr$+m/Ah</-ڱݘ0yϕJo y4:9qz[r5v'N8ߒT:6 '0C3mA 4 -:! N1 }E"msfPhcKKH.8]F Sޚ} N?6dpx#~kZ 5cVP*tDD4jNZVguJ>(fzr[`si,ZZjԺM~=M(0Mn $~=i,yv*&;n#6T*Hz@ A$@ u1֏-kYkҳ)%B\d~f[JS ʰp+YajI̳m;趎S jK44LW() ƓF)RcN̾w'֤XnXoYȖe1 &[Hp-dYōP9hSYN1&RU{-n9"G;KsLxN޶V[061ש%>ՇsYoXJoN#b. B}[Kʈ6Nf+-X1XB# q#'GO]ۉzabz0 bGX gH-_Ww#\EJ س㻂jaPT[ ɜQpur& LnA5" $.N&өE Zgo<ue賈5W-}e5%X@X#@UyEԋpzSa!,dak^]V˘th`[ \OiZi|Ӭ{ ^lOki<977gT!{09N{lPڠuބ`Wq4"vrt,K~~ DH-p%MXCͮQyYL38@&^hW7NP1ތO.QE31z1{PC텱 AJl{sWeT"N@KQ~;,a]pO~5?H>" -;M~WWR =A?<G{fǯڽ^O%cH}Rm>68!2&nd֦dr[$4wU[Ǿ%_֩!Xul{F{Q"_gtRkslx1=hR Y*rlLlEgY,r(PQăK#(zx? [ɇD|((,٣ G6 1% SɇQQwJzlC5#1~ S9.U!#,FeÀ\Z:uvO'u%e}N&14^jԘ׺`=&O~: ȮRŊȀcTVRrH)לig>&,\SSjc=ZG7 { ӏ)"$|+eWc%8_6jCr)d!7#`D\ Rp h=eeY CևnūPt݃_q-eAL,?ꡔFGZV` @T)NdXK~Al85,8UZ5$+{9Ʀ$09!dp20GR|jb@sܝ(I,Ak 8@^7aJrml$^VFWDH8:HCcumgORˈmBL7E*H= !C?ie{yʪv(6<,ڤR Gۂ6Vc*mkXaLڤF&7-ْSɵ$]JՌAjn}Cb(~8>YILPZRo]z3$F/)YlV85ѱ=Y^y Ƶ.^l3W^¯ =x|]_ cz;:<4=]_ƾm_ḷx|0g-hƩ`oTy s>4>$j^fY+[ 򫆹$}~sOgC[sP[|_űx zXkޤ-ݔX! qoS6\ V[jbTI &nvY:RL)P#uD%FQ-VᬸgWL _~!VT 7>|{\G0: Ee ,p#>| BSB^ R`kr' 0 sb>g h-c*( d!pE1]ߩ4 ~ u JT0H. 2C >pY)ry"%a=HekG2(f)C\gJF/0Ii*ÿSosWnnz0X Kf$26bX ~[Ě)HW 4B`6{%$vɠp[aQ W l1= ;=_@̐xe[sqombYKlN4񬝁dImI}8ļv8&x쇰} Lgz;ikX܊34lAMqZM|Reqz,.Bq8!+"ע]{ҳ cș4+ad;bߍyCԋGoX>iW B8?;= G0~fm13;2Iê/155&ղ(O9uJA,\ X곆SKE8g<1fV ߧ/d I 5w,@s1Ĺ^>ۇ̓>Ԯ$\8s=Bp?yIӦޫtx`VR1Gn:T \Jw#ԯPāyPLZ.#vP̥TfK 3z żPrFoҩ>*=Ľj>y3 { rJ"Ǻ[A޽fTW4"OQlxﲌ{E6o+x#8JIW㋿^ah-1fiT2f<4{ڸkƹӊZEZء!3&5w5F|h>׫öTǿ_goӮ-YtH+g~g|&]EcncV7Ow½%5q |Rʨy6G ~Ѫh/ =u\"xr4h|L!,t VpB?H5I})²n}'XrFi*XZd+zJ+b%>@v-C:S\zμh> CFlb:Nm R*Rq.pkI-x¶mo6 Qhpi` n@9N;Kxfk\L$Q7hA-іCb_|,F6'(:X=A oϔ }'ӑh۵3s5Fk5s{A o av?8b$hdS{~T+Ɇ6dwk3f.‚;ؐݥ ͠]30sN< m}K-?]6%:psjDpQ6̶UoImCeQRTb[OcpKY9(q&M̢c8BN $oX!2.=-y+5[gBV!QJlTB#$~'Kj+_L9Ĝh+_돘 OOhFڻs@N~jfVkͬhv5.r[qۻ.0Kڻ}R@e6_m!˧0'lpU躿u;0a>O\Wi`g XA<9`xqj ܰE[[ jr[L \HxINX7qk< Q5|ocjhհ/AK+ͺXZB<Ϯ7qh.3DZ&Kg M5o]s1 4%~J YzTؾ瓤If l`MD9Fӭ \sVѭ`!S6њ fagXP~Ckv Ob F{#6,&å5z ir `gOZc*ҭvq#P&t||n!xⶅV U'{ LƊL'eͭmw%"C !PϚ\q=TW0q-C7F5MQ&}[WNAgw|@sCn'L&۬2+/li4-i0MzdJ6Yq9sM!0u \WN-tY۞mQ"UR 'IxՎTDSHqif4ص1S\q.ekv p* +VMڛ0,l{ SbJfʗ* 0n/US5 \%uBY`/<S9齘A_js=QFpXPDHºRjνN3;Rh8ߝ8bUtW. }uJwbccWG.vC1^b=:NvId MC%6n?f'O'܉;ӟ}j )N\Hn~>oekkE03|M&ho :Ⱦ:*-\LmSkUj:nrtH[*Xr 9d9Vr6Dj! ux@݂+w[r*nKSREo㌤?JgM B!HL2X" -eL]ĭY`k-{(KV dHV>FiS"u{cj|3h3MLY=^j7LvZ0=wǃ]dSb.&y& g{~&θ c^Z^gG3n \3QškTL}0HK@шp(TD궢/&8]:5Up#3„oy=1.4;#I ;LW_jc[% lgJ.QL0.({΁3G*u ӈܓ^-tVa&T| F,~̶  Uny.*i2Sn (c۔Ͽ}[5`LL)BgsA´,җ@cZ|l+ȿs54b ,~͆ќr< ~dO°ԕd`PC"(njLJ辗 ıf5؋bfQo۴e}> ~b y6.[Sx:*9F͠X x+ȻL2#n(|=:SZ4!@-7ˇA2`>lh}&M@#۴,'M6k6 =;;d*`t\y&kne2Gi s%`}ڼczQ0h ;/s(>au9<7\Ok)@,i?CϱSMN,OHxba#/JX")}^)HoB)q]Is9+ODq"o}j-ɋܭ/?P.[TJNUVyS/Jg$x@mQ1N-uF9[rC޵̂B?2{h!(0$o`F>:5CA!gI^ >Eo_Yڭid P\v?j l1rشvB>19t60k>J{uJ@siŞlSM:R aN݌{E/&Cb%ܺŪ:(fTrzhb1 Nh "x 6:MiڌNft6Yٛyeq j#RU8]ҡd/l?쩘\z+CzoQ- B[ǘR ƺoivB#%9sM[I* Sѻe*~wMttx/Ύݜ 1Ű@b`;y/ 玢9M8dTBo 9sMEJN9P^ͱ~A;T=x|*Pz*7|E toh৤K}nl&J $+ AaA#UZX)wN=xgS Ee)ܮβ&h荃s|zeV6hA]O۠vؠݗnt"9[nND. ="h77WўirfCa_ɏ#g"w2{"ݍح޿؍.8q3 jF6t!Kul/Y]n;ti.ubǦ^}TFw+Y~뻎xO3WY^iZ+E3;cL3;?^k W久iL#Gȣ{m#>uq#y-w}*|䝔-wa:%O zc^S,HFf-5gv,H~籒yvmܠݏt:&n>vehp۠ 7ڍc}bg#9۠q}tߚOܜ+@vsY%^Neݎ!m7G!Bvv@$Gf0L;=Ix~a0@Xȼd97k ;  j&y+&Ŗjн%"C :CE]wݨ bK}q&C_6P@n>0ʀmX]yw0*9RdJ^TSaތP% @ E}mqk F]0dm􉡖u%WBe WS?Yqr隇y[T:'TU%%ez"K, ¢MQ95bNnR|*0 TK<(\#ǔ| hMA'W_.yi9|[:i{O:[Bk"]Bv$DvQϱu,75u55ԋ˧npuvz~ФQF0̨ 3Sc[r~*ag}Ԇ)u_Š@ NM!*#k(@Ҕt&L#a@ky4dQ9E:@N+-V"(\zϑE_vX8\bv3#:ڂ?J:',yނ[yֿ!nA>茎1zkSLTmX{ކW\P*rOر& Id{m5-5>|۾Q;ή=1h^]ebc㋞^^tB(ŤMw4+|,}`JP}ʎ"2Ub!8*/3yjU6t/-~玷zƍ*0%=ݼ)=Lx(1ע $_'kF5W,Y BCйIy]=fSSLt{lѓG/J b(%[=)r~0?}aOİ)oyz/hŞ~ *{'s>`['-tR͵2N˗/'9cE2  O:՛7n-ʸM5+DAɩ\ 6e)BT p7)z'_9&{=e9RZ.4j :Q'6X]T RBﺳV"dҨ uJBn1*]1KS 8r.iQB8jMs7tT%NB:J6^ġ5T0Jӕ[VXfk Z2WEyos2t>*az+dkq/o-u!'=3*;z]63CDweǮ(Pk!";cO8f*!շAB%gF+l*|҇?~/F[Ǹ j=t6E<8FP9"Zfw}03z.i א#\2(x+AQG:!.{'f˝(nsrL]1J*rrY/AskEd5G'E^rWS̖V5xuEx_8Y SŢ;=Zq9;|'Hh92f&ietZKSVY:eFv ٌ&,UKDp5 s+ ?ta~e!=?bеʉVΎ3$pEVGivݼ9K'mo1tsO(x59؈0֜mQq tI6䖢eGwޠ0sP|eO }}>b=f$s|v6:urޣ~fvݤX)SK,WS N n2jU+n %nPVYV7[~I;wnԍ<V{'{Pq]{v "7'ܼh?Nq>,ozVG[O|(!<=~ h-`K>K|}֞~6]*5u),ζLj 4ߌu],;Ф^{EJO[TDѓlYe˘ؑy#Ɏ3Kʡ:,W,)1?2M )V{qj(M@:4Lp LTL.VOw#gV)BJd_;gh:huJ=QOL,nYKT^|*@9ϧy52H9`jKB4r9|:<ͅLӒuZԹiMu ƌ r{ғeuM-LZzt[CJĦ6:tSaS]*@2O!t:Y k ?]s,,C-TfR7=)}rz2/]i_y!?__F@J? Ռ{gjt.a\E;m|.ͥc;ܿg*##}ǓW$I/ 좤<00h4/; 3;Q,Kdv7:XTH1 7ܒJ̈/#ݽi/-g`*/*12pBdIՀ4e/Zp=WC (0:7RqK.:&ԢBr,c\O1a)E^]g, 65߾_lA[_‹k Qp yoX{{C̗y+{znO jA T sf3=wVeZL|ɯN3].`>}n-;sk@'&6˸ʥݒV/LFЅ 3Y,";WͦFw}r}u+&yНw f4a+~zou|ϫOi}PG頹r`Z ѡʃ&'񲾺9ibBw}SY=/GmXkM"w;'s&j=FRU)&*@Wh6qr-AvcCG ,2Y?_fOZ4&ph?"&u=dG@$5_QP;͔3 ,axf.::[}d%=>FbtFƓAk:ycPa*weE1̶%`@)`t DvڸF<)`֍o(9F94v#ض (1'(9F[Yv~ࣵ:uTi=vN128P> &2}'0ZNVF̀={zR5ى6xemv`6A;gcK)`*Ʈdx G&Q;N}ƽXŒ ΈږOcbZ51U%sc NZ^3Ӵ*i)ׅ y_)j,*_C]RM]==׸ʋ13݆O@-ĥku~눓(ou}O2P*-谣E/!ѩA٢`qH gILdȈU[wWSEj)#Vz}XKS=, 4e`U9 6UE$ujkg_ >oڐ`ծX;'6v,xf?<=>9ώgB+5P(Dɤl}9}%+kȝ-*8*o dSba,!4M. U)sj²@U}ɤR19&㡉\G| Zi6iqvXW:Z_;f=M-1=MpvEVKyś YBCM8RQ 2PUyAϮIYKk9a6I%8'A֚;E]Q70߹|x<;S0-8J!4b%yG&˓{"<ǟ\׭CYTg]bA{ЍY]<en ER'gT*&XlGdxuL*nFtnT[0''9N%㓙@w86Sחl۝&j}7/N.L73+=?/SVYwu~^<AeP߭!s%vKemMfX|Z_9x *ᗸM..o#{"2w~y(>/f/C5v=~3|'窹󬺈wgyb|Xs+ 57bsjPo5k*&,TL =.dЀrgةN?yj~Oˬ:zIQY8 E;'|F7ٓ7?Ȼi ϛ;usfI\Kb)]`>ty6MdK"[:.,,/ԗ0OEmyO޾7& 8Р=&S#hb2Al'6 Fq[$2da]74X֨d >U'SgfU@# bU 蒇"&Y\m#׷/uqt`㠅X}.jji&JnިjR%?kT=6ӬqjP\9^;z0Dyaq P}Zz)u#aUaaXˡ+9XaSـv3*%w$wuUUm|ӥA<֍8ZqjT:nzZ޵ vB'& S9ә>\zKF[ "Wt7jIF9̛M[;@OӖLA&'uNЉ)VC`٪uP%q y` B"o!Aݺ^zjudg?.}ovY/eyEW߫kOY_vOuvGT_&']-~@Ͱ~4c>^}~_Z,Na &"+` 7YI,Tvڂ\=eZmtcLj ԿyqvVbڢ`-+VF1ኜؒXe%TEW. :Iy.KK?}+ח!y nٷtKEs>㍨;5%}Q;ڧ錭nsP6.4tn>٧4#D B cTMb9e@~v̝cSmapuN D*2ethM ny/ez]Lk)}f g\1. h̻P8.]eMbyRαǤ72K @,l(}5uG_1_/oάm{Dþm8-wh^S$1LG|W*z_ꏏ`_'|(]>?8EG䢚BG͸oe2H&Ίn2:%`yn1o,$tv>Z^d`lJ).6JUP^Xdk X4A2Gu]Cɻm2x&(x&)hO2?MՈ4ǑYso2+M]Ɏmv_mSWL s{u)7xո9>eOEű/–{lwa>6"“Ž>aE+_-[@KȚ}Ӊl^X O(!eumIV*b&Ri:X5ϽRMɔ o /%nEx*2!ՐQY5J0@B[^}>a){mu]/Z+q|owRќx"֭Ely}G_4V}Kuz6I#F}Ӎ;g0F/{tCΘuU?XPR+Ÿ@ޖh`؛iȭ8N^G5;P ̥6ff a^M9go7n4'lݍAڻ7bѭv3|v!Kfݱn'@1)";XbT-Cl;r[gTz/jkA.8 vj$X FeAjw6CN,<(?0ݝ;?¦,H݇36\?dQ9 )řdE)HF#"x-_@VUeKmG^[Kyעk{k6C(Iai?m բrMT (5NٽUta DWyqnIej2oѯVԺֿzwW_}zݝ\uz+f~*cb݈ZШ5/Ѯ$…{@‹bF&ݮP!D%u=1=fl~K&kD)`嫏]{\T ;"6A+i꽓p0D;TLQ5E75⽋N{o͊yo`RJ %N_] sJ%mHS9L3ƴfRG|0DJZ 9ɠ8페0d|ݻ9'QW;4҉.Tڬ3fq@ʮAyk0##CxmF>?l8o;*P:B ("F^W/Z|}Z, 2z3̔21sB+fŅބ].gc0҇RS_\8@uEGLlzgf&tgB6-Gр I"(*f#2&4(*>GQ+iW$nɃ5 y(杂ם@otWpAU>_~?a<a2gч`Stw˸UTmͳVu&2h8.ԁ}{JEq0l刺9,N3u包U7=ڬ2wcpF5XGj @óhݜ ,L}o !a>$GGE4fzH6g)һK'ƒ[kPQ`"9CB[7IcGX #WT5 3RQ-!GKȗXb&~gɶ;Z^m]]ls8GQH zj&oRKXNj]/4Rm @KQyMtʬYiuŏߝU%G]PҢ̦_ Hjvz>k/1&g91tV: טS*\|8fl:02 K"8xkFG#7*I8#z{.Q``spj"Xe21HZ=3V#f)+ :o.XH`W@(9#g pZ<18!1F ʵ ) HNH0-'.g -G /"!v[.BRq/)QHģj8$p~$D(<&e(`] %x+kkF-⚾|LS,(YrQ?mJKN3.;mYjY_%?YGO8CiCa 3xBHƘwr+{hح; !k=Hv삘saJ/)$,l(I+nM٦ۗ<'01}A.^LoW.x>] s3k$J-ا%z=va[Zthr%Խ_ȅI\6DS%VQ5]);} QfΌbZQ>F)TyWvONO7Md w9w'UӺQ8$pgɒtJ+HVT sdƻ95Nyj0&c PG3AAmG,^򞹥e?;=xR0X&*P{ɣ5tljET2Fs4iIU)DVd{;y*蝿*@d)=UDG;)V2bʫM=:)_._1/6TlU{ \/7~F9PQ+T_G!WUd]PYx\ݬL4$L_>΁]UF~wI[8go1}6/E![$Q7;f,=JyFaILw#iW[4o *h6Ws|(/הʼnR"n.`Zl}r BKüZ)Lw֦|o!!ǨC/Zdtc9{*0Ժ{$ Fu|N{}H$.D 3f% 0c1ˌMz`qS1h2GBb*UcbcH9[\恤eNњ?hݎ=-=Fw2C(qԓ4FS$\݂z w0߂+$?W9X @eytc̕8FW+7\mMy}Vُ⭗Vޫzgeu8\?FYݍ|;AWW^Ӌk77c_3jͩI\LKRyQ#A@Pg,O)B`^/^< (?NX h0Ԋd娾nk K>W(dW<AT˛pSX8]o=3:W\$U:QΖiF|8--qZg)p\y Q|yF0$,HJ"yo2[>i{ɸ}2M]>7MfA)g ~5?\%6LrW=k],V [:O5|`^Cs|Hӵ`^=}J&{˞>@PfE-0_°o߉?Mr :q f0|gxj-yOb8eBKeVyUU:^bX h?OXX΀FпEoy-<)Hf'AOMI\|)@-Qk=io)DҚUL͕[[]+Wn~I0qB/qlsvXqީ|J[ J]AJ{̽9)G(Os [0!Iwqcaa3 aye5S\]3GyaJΩ=N&O޽.(T,| I'{z|Zŕ&;^Q6?6KٰhU{,g $=#)+oV@a]~HoΣ-{769%3{(sB+fa97a(*iz6NHwln[c2iEvKrT:֦hW(QGICHD5 k&7}M#+j}.A n=t^cdn/!Tcl'>xxCeBT,Y GѥEq?M=3qui!c *2g )-YpQ!(qiM~rDͬqnl/82Tlu%SKe?ZZ'A,WRLk1#i_/#|McM ikˤd@V8Hq=CRڔ=TUWu;cia(A6 gUp![Fk| T"63^??W5A4G~xS4kG`gϸ ֊롺,>CN5.eXaKN"5qE.d=I.iB@YI NIPU#2yJȑ~vׯ>#79~׃_|5:gu{&(3XO仇xtn6'yp\JfV5 ~f!ʴ"0/xuns;D9?Ϫ+IpYm=y3('q g.GA"Q}yW*1|W*3]jܯ&r椔 &J8K:ꕇ?S.@^͍<fVCKQXC>v9|j dgWaj,3s?}ZKՍ;K}9#ЎAo:iGbImz7=lxጮ(>ឥ޾c}6?DW1XB{o狡>0cL5Pv%%hw[Ć0m _0yipקX(BwܹtAӎ &kuԿ{9>nƽuY9& 6_d!e>RKٴg1_5ӻ:7}C)5N2 Mv߻w,W,@mn򒦇kPZ?F;~z{POzTaqͧ/}`#)JJ>L Z H?bM=nҷf}紥Z1š>xuFkVcT 'Z^cLIvd;_kg(V6*[be|d4+Cyڼ*I?ߘڔ?z/T0vp7¹ܷP["Z} zp|}AbK[LWI4z:l_i F)W3bz+=ɆUWiLS_9*?忲I}Bc"f:|>Vv[;+QǻvNsi=m:]UZPeH;TчyH-%⃦2II9IࢉA'?|?]ΦJekTߕ Yifoq1,|cYDs(8tbD͍Rrh350w` :Ù*lΠ2ұJ*PU<f9hOvx|~控>8n2Xɜ9!)EӲ2PT+\{e{|MFb ,\ɥDP+bD`P/d%FiՏ/],Z G7oV7+"F凡Xl)ޠJܺ&+x2U BԵ* Ũ NB':Hh1"UQ_1m/r4Wc30f3%흁RT7ݵ.R826S[v@Mzoh%TaRV +D"@unM Wӛh鞾-3\mK4%2t%6hbHؿ8">ʏ-A/[`.S>jƝ׵1%JPZ!cN%Epwhv-N*;GuU!T ISJA!/C  !)bL$)"A8c$umx}+HM髸8|F[WP I`r-{ eTJgV ɨ9*Fcτ4r]BE~MQ11踇ú0ynRQ_OQ*<*i4 A$3q B%x(s8xZK!ƁTlR~Qˆ*`<A !x8\$Dj,fG"ETQz.QQ4;dŖiak#SJVY.u%}d&$p-`NlJP3DC{R.OHptx+'gFO$btyED4`fZTHE&)Ҽ *И].|y?Ӕd|ƛ$s N͔~g_s7k*j*c=h<mggvq#LUWU_oo^_4LS~as;jO5ՔhM4 >~H}_0n#j NИ|x{:>2=(!Sn>k"fL~57< >3 j%Ty:AJv~qN!Bt}cw0{Jϯ:s 7?Yŀ]JOair$Er4CO6DFAa3LV=HaiAcNUDI6y 9%7F1 8P=h2'ȥ̓Dxe ,2 2&<.^@ϳQ#­%/&eOR0R`JCUs}%($g1ߵZʼQXDCE=KQ\lӞgF*$,YR2V5%21epM^6*d9 hj .hR&ah{=z2]ܟ^E_^LcpvLTq 7W_1I湣I1F(Aukmp[RFK8m")*LT ~#*e$1E{XA@eml4FGM)*()[Bŋlʢ9T»Ib>J)j)(E *yū+YSƭ(E!Ϣ(C\\D jcGXa;\.d2ef H&&%B¡"*T %P&h [k "nYQ:]&虰h6$ӧXe^+'i&˲Hxv ƗhZc= ɀP.5dAZ)UQᘨ4niA9ZCc$ڔ^V|@k'α-ŗpZTa?8~dg/_ߞ:) .^\}y2Fve(82eEqیo>}=@`wkxJ;N4 YXd/BR(v5zj{ 3k')[Ύ0Q&OwE PFdE-ڬٺ,bR*wAWΞ/jU Zwzk3et_r˲ʏzIERsI=m]+iwً }ݵ n.d$:oWyKLGj;NsK7l}F$w[,2t5kyդ ׽òq1Epw$lг;_Ӧ%pIèbi;8S{R`##X?K!tQ.| `uNM aQDZ5ӝ>XLkwSȃ\d=A\.]zzak1KZb]O1w {< >.{|wDIϯN̞n`q%r?U()YQ>hKB1hf1!ւNJE*++e55&9wbB.Gh6"w7ʜ)8>8(:j8hklZrom.9'M G[o-ÔpbRU@@koKxQ<gtC.SZUPAEh3H r Iȣ2cƌ:fTu=ԅ^@e%x.tlV L70А υ2$jL$R&6'$Iʰ vY%6Q'>Z~,`0& iϼr*ckkEq?XHXL[D9BSn6bn߸D~16*X,;x(20V"{,"ÈxToID=.s8"!B"+|vbh9b .>T_.1z#X4V{=IVٌ5;(:LRS2wKPmy'px#,cg- JUpJxg'\<[07QFxXl$9&&o'tZ-c=d_7qǯl!XZcD5#cu 7c|<y^<go* r?u ޭ,k{Ad!$:DٰRNH{MoGOv>P Z큾GSds7:PQU6{l]\Fx c[``o18$FY έdd *fѵ12j'f_)UP`œ*b9Eȑ猒 Feݩ(FB FJU>$C 4JRD KtQ%QH9BW=YnvoV|=T8DQ3abU('Sە>ms.oߚ; Q Kg;D}dT2DnI;G=" !r; e `.+Q4T瑢JRm#rڶiN]H|uY ) bBT6^B4S׫i=ͽC挼6`˝@0n֗H8a:9qQAޜKФ5eM-lqK7H4A+ a:" {Luy(TK2U+QSVWJNzz3ym'qyWyYxB9!;uƹܐ\pN7ܽB` uqfÚxrGSs%cƴ_]p8m1lUB=*:y{c|,Чٹ ޟ+{G%JDL1m0jVPc0/"1s)D*0o߽ݥŞIyl}1F;˦5A4ȶOmuNY-U'ymn>[Ii}PkCnc8 nG]48I}h\ }SN- \7N~Ag+R(dDF4[8*9O4 sZqcxծD\8lH` 9pXId5}5w7>wDE$*PG߯,ݍh^ 1~Fyƪ.A\gy>ʒ&,Q:[^KzF}S "E1o!&(YDLg:4\%(2r}Gc.Pv$Sftٻ|#'sѺz4ʾj=Uȇ֖NuWn(J뛬NhR,+PՖm2GTϜFkedTt|Pu2Apte+F>j_k\5N✑_PZ{Z)+QѸiš1V3Vuv_b bQ~k%υ֒޲yxHۘc;#7eÎ4.T{)ffnHaqAqg4_d-;&),__- zgu\Л]xiĻz5\uFX㊴6 ^eNZbU{K~ LKp0G4LYK`h:otgy\;{6$DoLWD wCK@oo|*~$70 GfF7۱G:z2pwBFi4~wΦ/ifLVqv;/D}&m 9o_/'Iڷߧ g 0;S/0& l(("8~?1z3:O9zų1?~g{k81iuĕ>2x`va6ftA ?˗|F?ǁPF0=mb?V0co*;.0 `I|Z g a<2g|DYE6*kz? 畽O;d-w0[-`vvI@lWo##Kr8"Qwp69ۛy~^%|R`?' HM$2N~"c=Lik]}G껒&l]}utݟS}7;}"qe؀{JIcZ!&"Hg[ɝʳ’.)(Յß0X9c[yz8B,:\ -oq1i[\ݸބz ;p=C*tH=18=JT@d%1)"~ܥ<+,/ю]&s+\hYpFPNBRṜGnjA[\ǎ'tqsQAb?g =q1O82~4F>X_d H;+$.%D#ȝv JRE;-l;d)x"IH1/PX֊y,4mr"h].~y?u.BW;<bx;9|I/"fR<`s[<7;+$.)) %Fwb F۔;,/Knέ? /:,)k/㨏:b-gn S@8M`Jg{f++_NgXrmPW~qK 6 s !32qClw#D!f6ѷ vb:Gew~: 1װـz4k34j4)dMf ~z4?{Լp qulm*ɉkd_6"ОD;I忟9qn8 9G|W9ZU0>CGSdZ: Tf-ʇlBջ:8A&9UA(.|2rz[SrP޽7 2YM01|/P-{eYuKg4 'O|=}9Ag4Z|zEh,'rPJ"PO(8L `"/5Vh/2ϕ%l;}Iް#qH BXײ:h/v~:m+X]LT{UW}}a Zgi=i@h5 "Z/8CϼJBGƚe%Z(J*Ks&eUwZ9v%4`,cZꋒԗsǗ܁Jӝa6\7:bj@)j,G4˹w?1(8E5hBX&ੰL\?ư?9~A"Lk'Z ͝! US%JBx4a9)$$C_~:we,8Yp.s*t;o&j- kEu:cJhVtdF4HX `aln9*8[*G?NQ2>+'>OXFϭ7 ۮ\珂[~t}c3.S,QKJDKK8> ]8\R #Ky)8I0 PaCʒCOEf xEz.tYngd6hWQwx{@q($MA2Q+-r-A6 m ڬi*^ZҗDVE Q F*? ދ׳i!!CvW|/ۺC]kۯDUb*Tܿ|x`Q } _dndFp}5R$'ӐeU&4s/nΉbsD߷|U̥=շ'N4̘ u1W0Se$07԰g)('O3q3Zq624 mh1 !h)PTD![\Lnj>Upm@e5Wp ^TK'pJ7C."MXTsW}#g0B3MvaAĴecJ R.6E20sU,uʸal'&aʀҦׯ9Qхs~\߶Y5{R\6s9% ?7% v%W8}(wo=b1wrǵ ٸI|mIS{T fR]CsRs7'=Bˋj+Y5U?Rk}Z1=$+8rm+K1r8mZS7tEE=xT>ƶ";QOɚ7ׯ/?^Lw$ȪC@;^&^'ғgɗnz6ys~n~Q~ϊO[|2Z@.\T=(BSrEmkm dt m5Q6LjneF}#кqfEN2Dg20DgoOi~zvC X=S?M;$YO{lE5L"ulD'P| t`]+fTE 4A{At |TΦګ|\ m ( eU뢡fowW)&Ey &4Lvk]djo}thx"Z{hp%Pt]lPQNwL\_*x7$@e2$C' uˢLBϢ7T\IyQ-/_jn(WĪUpOj\߼B~=! hB(p|}ly1fFk<ȅѬ;޵]dl^rµ1. !̌3^fE UY0RP(-$2AzUT{8E',7Z_uitF? !:9Bc}$/|l+T>6Z s3Bؤ{zϜyZZuF6qRge^j\2eNc!p^N2Ch8P >BJT^ &n )Ad-PiyJcavs%0V; B1\%E)J3ꮸbp̨aE4E  +n}bU'S9`bOq;gf jfooRpB #k* Dbc8:L٦ˊ3*lw)}TX؛ڶ(#}Y9ó=ݜ-ZM"-Sj :eVX3WZtWH}x Њ6׮&B`Ӻ /t.P'] 6=}1]"eP>EO I愵bD1STy-5reVdVb!LwQlDc^Ri3P.B~ϴ96H'גO^4,+|䟫4M|s`ڛXzԿL*F2f9z~w?f=3_"nwy,Yϳ&.G(YxI@ZǭCN+UAp^@;-h,R* Sټԗ K& y[Q%]/Zew*N}2wo ԛ~3kt_40B$yZy&P*Uq_$D f (~xd^Uo4Ӽ}?{*/rL~ҡzIрA!|LgB1k :BAJ{\m|}sAC}9X`e Ps{ā6Ӷ{dwԢKXkr22hGZQ :6j Hnm$9cZVQA+%켴tLH_ Lk#ZMȀ0Nقyg1`\Pq\:iJ+$m0+ P=$sI7n1=c(f}ituІ6")eUJzJTAh&:,A0Omp20-PѡvuAL5AZu,3iP;T+&ҵ} e?j/b+K mHŭwO aDlMMSK6qGag趋ۛ6qq)->f վ#mpaR8"8[|$Z͕h-Д@"U/kpl =a; RWRL%g TcXhԶK^hD*NUG )X&ʟ])\LWuPq~HǖWu~i I¬i@aח!1+o|:nʥ?o,i]|uIaxS zaTobFw9x}~X̋gY㸍+[sUfƫ8%}?TxQ$CR{ gwgÔX` !P!G}DCl4ply0hq>CG +ac@ʻalkt|Zpkx7( Wy|iTL2g˹kV<]Δ#{zg<ѪSaGw}R~{ 928kdV&h#ȓ]eԪٿEwuQ# D]Ƥ'U4.-"'dhV|S>WEkJgUp;^>%U \O*GeGK{iµ/wi.5ZnvZ_&;Z~~\"w&2h59'DRK\{:.HsNf Z`x[.8|IBǻ'Nw7oM9dFJ,KoʸӇ,3wR.X>0D7?r':I.*.͟rU1~u2*7*~~;:0^w,kU(I. JLgRŶS+Vx'ct9L9=-&[Q .FY`@)Cw .thV,zr^s{_ߔoEL+je$ӗ?6 _4yW̞wyW̞wmg1A9j.y΃V(94\[nU`u`lϼL ^py 4hr~;|Sj*5 }dQE[qVk,VâYZ||F[S̍>1==RV kdLmfYsJSypNg&SFPW>|iP0fccvnZ5zۦHs]Bx-WŬ rk;@N! ]!pJ4ckƫ{)Z\[ހCMA6 U#lb7Kt sA"ݜQ3F HR< ڪZO4Y ,^5O'__^!E儩S%OE8}ץ3zFM@JzFCHUe܊ Q#gl*)$enWOSKm(Wj႓>X\N{Cog q`P0)'p4Nݚ71 06A Ac#2gj 5 &x0^) 3GrP :*\ <꛶`<"7؂U-8n BYz.EZT؝s77k۳f{# 9F}hG%rK 6)'Na\Aj p~s(=N4q襏CHAhMxT$)ׄ9 $MŨDp^;"b>fƪ]ћq8aMgnWֶEv'en*kkNE˼$ʆa\e 䪧q:Hzu6xcX`Y" mwE $lLFnd$+lRjfnr!+Lok ңRLѪr]WT;k=Έ)} kLgXt3B -Li%څTHR@8J ~)[˖Cp+{-*Ia@y3($N՛Ѹ9f8OL0Zޟ^@4NѴO.~?\N) ~kcIA̓fqԠʣνAAH(vC+β)5g͖w,Sbkhe j2Ec9?MV0?|Ӂ?N<]|SШFo}ҳD&(2҈sC!$Im֊C !smE&r91vF)Z~,;x#ύTl)F0F\: 2ҿB>Iu2zX&M`,ۜT\;'rFc?(<^O?M vE#sInF3Lh£7.ybN0%5T,FD f ) %RID z),bAR>nnM/C}ci{C\ӳm"i> k=`17`ld$rǀhJZ$/l$0Ɲ:C*["AAɲ $k65X{6䉩D'Ġ2\<9L d@jfȔjE=RpU"̪<2& /.B)%|j5 h̀*2X1BIROt8%~1_ۮ92%f]BF=~;ovٵje,fn~Wtx7Z'NSŘ 3kkrIm>nmrWlwQ 'zAaJ,ֶdS&h2CPhk9\mB)yRHDy8DbYr@U_ADmǔ 8mJ/'hj3Ƿq12k׎k)Fث U}mb1/\5G4 &n a#N>@Zv(L=3rK0ƣ:rhhuCː(zq_\W =/cz a/?qgd c4;;=O8h뙜9ak(VٍǠ+ iU h- .arfʑs/0m8]@8dɄ̂9Db甕Al<(0FqII͚.Ⱉ!V?aBu{3ܫBiپpeٺ7Wb1 D=;m4ecDl+b+ieę1mv  s2e\٬+hVR$b+ klz,j,r X4hKd6?[}*99圐ݣp)KX6S,JĖ ]Ldu钛OeaoRx[kAjD@vKy4X72RMB\A7M|G'ZːF=ܟ@C@ze+BH;A@C*ebCK"r漹"6F\8impfh$k'KiK58k Ngd5?i}<6 /wۖ-bEE144녢4_ܠ-VNkǺzyF{"RbΞqHCc3jvPʲ)݊>Bܶ[Sfڒ Z{GNfj-G:P_1Lce:H DrM\R(Pe&+oasv9NIٚ<('>d؍rsw#~(6jBoŪXpշ7F59lK.¬O=F_]i ? ˒Icml !/\y < mv{ Ϫ"$ikΔg J" 3>Īk(%8uakAd( (Aczq; +Eo{#$G{ex_0¶|w_so0FӚ~sU7NVjO>bōq[Qp%*IJ;$Fٟ0ɏPSW$lMy,?Q'joGyxrP!ޝ҇Z1,YEGb2F*AQ[S9r uadўȩ$rʘaȞ^Sp[w%{O3 1Q@HO4&'t'I&J/27B^\nqp0jŢ-#7w$Kdˀ%$I/g%) mu١$M(MrJ:d&Jinڞhk-!0 G5uT|xTA)hT5.Zh!],/+=|?^Onf/u1pRlޞFd=yysqz[erY{C8dKb2٤!{<˦Q+|HdxF:(0&E@7>h󋨗w;*=5 Y7*mb"R$됐I_KR05 *KIr*K#?%1)]҆bYxATKŠG~Lh#՚(%b$35) yG&0 GfjJ4l$3ii,KSd!ɿA1 VjfRSC?qPk@k)|80C"=K%M[^tHR2P{J4:k5̄=[)ftҏ%n}w=nD-0=,uڥ0\iB" LGѯލYRG@GwqH_np ~@`!2ȣQlȞh~DVĖͪ_X,VA(CD޴{lֿt/zMzuҟ;tn?S-.xd)xWzz3hg%Dx}|jCB.ZZԵV,ևwb|)1(7&_NZS{g|eI<\eDeɗݕvtz(F )P-Eh<%a <a(Dz_KVBߡJe9kyi3iXүoZp=t. ҉ í0'd%`vuZZ&`b%({{^s5n? w1+^ ܪmi $+[X0d]:ݱDdnCj]RĨe(Ab!FZ"\m(Au=w^!09R)d80Jm(GBrT;ɝb)2%O!b}U{֋Yl}vm̃Pgr'_ATH}s7fa_y~5!x%ʶ ?? Vː _MZz_U Al& E4I-v5hDʃI}G61~S%rH7.)2%)ǹRy#:hE L%rH7.eR$>{7F7 9hv-+n~V+6K6(! k%GnMs1]3.SƝoSO-.VG4'@D'3ņCqL10-3sϽ2r - S9ۯX(.kbT(x0}:;?Ve8ePa0p,\qJA5Z!D4,KhC^adrj[¬I -(޲e1Qㅥ/Dq,AYyc4zia  &̚VVch3_4ErbJeHk%YZJsRr/ +4 K"9 FCwW%}@ك~_Rˑ]> ObhJRBqIلPd u#,0@P p B8E@cXIWo`Ĝњo"!x"^8F#1߮Z cқs6) 2)T#(6ON{Pc\|8PmRQA b w&Q<e$^@s^Ucl2,ANC5h e ^rj@U2 *!ʏ)gL)dB0;^~eB[Q#X . p.P^X Q~YCD.P .;=.Pe<["@.!߸&TNOՒi7jA1["D'hAVwJhv !߸&T׬=}FkŸGn<wn#"< ݒ n9$B2N1|1b>FݪK0F$D=#QOJ!UFEQb @C ːCC;kw႔^e$/yIKR/~rl'a d)c~r񰲇AAKaTg{\ybɬUd ) L )IYh 5^f{$8BV,,19!ӄK?RI]Zt>Ir儬ڬr\ Y+ ~㝮:n'|GgJF@G'~w2CsR:j4I>X_rKUQz.̏'WnߨQ?1)4뵤иI-qݞ}ec~wN 1_q7nx*wpUFEq;&,/Tz66~|褙`r?]/QV]]ޝ@(G)X뎊$S Ʃ}y%պ@t-jxs #.5_{Dh݋v ն)Qqz1ܟO߀5x 09ذqφ1p7zoܢoGiQ5.ڈ֠ۿٚ܀ 4`Cn tfm\|:zǒd=DĢvLDG8%P1hW>`'=Q| W`Wd(d>VLpp71 gWkuWbTw%x˜$/2R)lfZ2a HR RA!,AI,J#R*0cұsr&h(0"3$ ;)-sWSs["'MmG&XD)u-%?t[D6Ix \A :Mb) ^#Q9H'D 4MbPHYz)ރ'zpHFa7m6p( ZPp=17}S2xJ.`*`ӭ|ehLIFl ڲ5߽{׃F7ܴX]~ow5F{ؼh@𤭽ޛkwǓ$j”/nMe~Zfǟ?ť/ooQLR`:WKW.)zmN NXje2 BJS@ EAq*=YH{]Ǵ0q+Q(&20  IITE6-@5/R\Mc! N1 %$f.qf{.DH&$Xn"0YWIk,ŕu$ےI8I^[,dR(UT %+D@rNCqa>}c 'W&b(l1Vz= hsZ LB"PPRA!v(Jέ;Fz 4IEI0jȣHcW`R~29[<' ֠O~RGSrgP*'۠bԌZed0 Œ}P1zdFL誘ɌKg2'=tc/<2|w $@Һ՛~ wSzO?TWV2\L*"h:9۾ԝ2qmoK0Sw\i%ed0_I5bP9i{U.,KDlh4y"HҩN\\ʡP!) S^0h ZgQ57nzqH4u YF/1I٤դb iB  )s5ՠyA 3"-\x% ҆w|B"'T<܇h/-(y}4]-O*9?YzXntuպd8@L>]@2ݔszKnPQGa E(ꑦnF:͟sT͗wgs^͝+w ؀fx{.~19?>^ٝOaQ7M_;su~>; z-@D /SDۤVg>3"iP ua[Ȥ-?wCPVƕ^J!3H1ڗj dSz )0xw{΋ȿ49eJN}tE=ZWl~.?T^~(NRd)sغ:~LYל?f]ŲsL.%;M>yWOG|zc/b@_W]zk%QkQ".P KFInL5]gڟFlH\y7ݩu?$){zJ A-TDnL,3n6Oך`DOEPX†"L!̢jFY0fdƼJg{_ыg{!X &FOE|#2k yB)#A3*>ÝNFwa"r\`8Y"`׀T[ v#%k}_ML^*UB >ཱྀ/Yp xXZ8EE``iB4QQMRRيnMr-S~1SIEA[D=5X#S֌hb:G\L ,({Ɛ–7 #E!]"j3E9{|ivUt !:jw.PG= N2j6Y+-p:P N *X՗;ȟZzPBuH&^S"F҈Qg<5P2R5P24Pj2$rwm$60!FCJ/"eҊ堲V"7^|a$z.ER"'3h1ѴZߥ_F齖-k[E&cS{$jsn%JJN P|й0'E?-r>$cƿpˡg<0uLDmX,`(qSk~ j4(1X$_*c6a0_̗WGg~hx*<\ !X죙Ǔw '1"4=b7;D֏<"9;q"/rYk.27ASTImlU/ifmՆSb2q`a&-V&܋SJ qM;55_^'eʲ^Dt|]G :xa]u3t\ųTd.v.im9\O<Gf6O+Ē[w[vo~zI5@upOW^7FuV\l\?x8|3t*ty%śc)*'>EN(v9c%$f͙*y…㔏yPrTG|BO"~S_Bg4#>1s6$Yox õW* 8|#H';[ wFLQ Y"j4j |[cAG ņ`8gЖL[`<@%!ؘwC8fRJKx@!HC!bPThEA+Ԇk T};j7-5L6|pρu&\tɀCEGZ=PUD"Ekꑜ4VSFI"͠?RT0$.z :)I_PeҎen+D.*+p{O??r:LzzaUǡ_Y'w^|ӪuO׬rC-u>/xSNj(c[콄=\ptZ>U B%'X bFÔ!kǃG*Ƃ 8 65Q<րB2W?(̇Wx7g?d@>j9:Ϸ+ZKΔ0ď5Fӎ\͏ uF`8:pA+a"ЈÌɅWaN"&*prZ"TᑪtD%6req Y!HM\`@Kc2K cXكlrۃi95Firc3?FhQh5*N{>¯_mvlC7. (Cpo։۴u-F)u ֭i,= ̬FH\v2\{;x>*p$<܀7 tG7}%$T`e a;ʤ0R`eVK+q@[B^?|Z M: z}'!>R=(ЗC%w*Q@n ȃ#$';_0݃&*Ԙ„/NG6|<@~S35'Fx(4U%']ZLV 8b:o(|B?QZslfx {21Ɠ[Y`!')zƣ[ o$k:$wBIʞJǒ-[ v1' >oWHL,Į{O&%`KOY~^ eK rbqE||f7M#'HH#H') LRpG=c)aK޴0n2}ѧ!64OM [4Z|ƟBpɱpo+B5uXTObVAR.^Z3n @)shI!R1Vc  (hoQyDH!I!$섁jt N2@(*U3eNpJb:1tQ'It1jWh8 jOUfĪ&꬜v0񘾚8CʻWTTXrG`NV\_3Jna| Twi%ZBh"Jnr"@ψJa-#c 03[5V7|1DFlCd(hHE@(H L2jSa:45Sg}>ep7E$VѨ}Diu*+AZc9k%⪚r&VuYvTVpf0͌FjǽRkFZ:(WbzAxJœGCqNLdUEE6PZ]Qת;?IJN簱Wby>6T%hUH uxœ*)PlU,@kUQn1Q=G2tTlv.eXnj2 ǏKgM "43are+'kΕRZ_U:ϼ}Fyei3 |37n6O *&&[R|bRˤw<6azHH R )̼%~%+q< RDrhyE-R*>ګQq!OnPeyp.[M*BroX O]q~&=?8|~խVWL?Y2O[M'V:<qse]v)|VRYVj,H$i[pǡibG&N=]-Z3MuUsfT81e22ZW,bɼRah)GUdܕ|OfE~`Zl .e 0f-ۜl3(|YDeٮb:́-Un0S7LKدLxozTDQ:Kzrs.] ޗtƱѻ;)ջ:9` 3in^n7.;:dj0m^w_}t*ww)FPiQn&!tʼnaj)$n-CmhW'0RKʖrV[BJO:e8/v4[z Q{YDl/o~2Xx2JTD%S܃g8nH1+T, ixYe# .?wCSESV=bq76wP|Ocax& ߣ!ޙƁEqQcȣYXE|=J{8iJd6Rߋfmk/bDͱKf C&3{[ٲ DpA_ƣ$q@J hE 1gS9.msb8&T)6b/l6Nyq7? Z@-bl[dcVBuJAhp[O~kN(g (:t-f |/]d#o-KPA[2م*Bex 1w|oKeA 8Cdː5.(Bdo.9hXw6=`O*'ˬ2S LG^WH]JxEJ&J Խ`e3aKT B ;!PYX+Q,eK t۸X @VۉfqUd%Q}M%ߚI$`tM̯YӪC]/(Ջ-f4FY'pBa}Q4a(oegoM,6c;8^38ނ(ι;o,\Vv9 8V\;u1B@')9Pb a|5 =E^_cJ4/mdNZd__jO.PY@PBQz1'E[$C{^.|2dQv{ O~ۻ\aaQ5yᆡ{G==gs4 jH[7 -凼xl|w~gQl ڃoBL*r0*5'H"?eB [ Bְ*U aڱؠnPV}/IҜf=!U+ [P%*P}ȸضLIU&SQ OV%mIwg^d. Δ9luJF,ׁ%bd)hTN$ r<OX)I +mV>M[!6X<^9L|oYq{b1/a _&9*?&UI͕$nc LWﮤV9XNW++DsDP d69Sf] 5 j 7 27zv^,ax2?,k0r1Wlku'ַֿw(l%-Ɵ'$?/f:H(Pߧ2ag -.(G*I=aC!')0Nbe`[m9+FOHk:N=@݆r4hl(}V{AƷP[b0 fuosj岶 0C2,(+fdK a.1G'6W+E)u8;%yݻc=g_؂65>@X]84 ۭ(`HB^X(Ќ@btiބa *GDw1qzґK"$P*:${@aB\GWrF)1ƳZĵ( 7>t0#1',:V1yA5a{s4. G@IUT tX>*M,͇W7?£b4qTM\G},ū@kIzu[}Dq=[Jc=ʉS~R4Jlxs^<ބ-ysܪr O3K0w2 @D\kK"\E˚gkB*~ImW[@ Ʉ-^ա+> ͻpX{4[w "T%N(":PSD01 82QZ"I%uvX57c3F< f2HDX)C ( pd^@Ip?-F68u̓~17]kEܽ< |W{1d:d>Q9fF!!8B K0աT"ʈ/(FJB| ݆3ahBɵ( (NTf dx F P1KQBV m\rFj~lXh.b?2Ԥ0뫸J`hLXRI$ dX>(|xĹblj(IBQ΢jP%:"? HECULVn<;dJ0q1oF)"aW,\q=TBJ O\#)NM/ A^1˅ƌKq2qkʮ0/%ކFBM8g:VX}hg HpC'mst4zۦ-\o]^-p9]:X<nG~A ?')(D0kƦ>^)o6|lVSC887 PGK*O%,i>aUe)O=;xǒ-#s= HsgŽ!E.'_y c"XdV(k0<%KN8Wb(b[#aZZƍuUSc\ tz8J},n{w39+oOa"%%PqV__mm~u}76Z+CKA*^U WUrtFa Q֊5k|_Ob׹MUy]pdRvoَ`m4*gӀG>sͳ?l~Pn]l@ M紼yl'<c}Ob6 v9s3Ix1^!:p^=B-g=5w mV5IFLd(]o[@I3ۛz³Y'wd( ̛vRL"' ܧFH`UaE ƛ7nڛ̦AhVjyl#ˠHG.}"8o^\ha'okVnK#) YR; ~L*Z[2BCOHwn!Zۏ[5C[ ^ Y *s. qbOsͷfgMI ;nbp/LrFP)Ccbae"u*Kpɹ gmwlGN՘"f+8:ivbgdG V͊Tvp%.6FsZ8֢JIE0TchYV*0*--|aGZQSlk{OLDG*'}Fg6SH_~W<#{I9kьiFc)po2W3smg.\X1Vn yw;jpbwVܫ0Z[t<x۳FP^7F~W,Q7P1<1 ̾5A9Íf?Lc"Ut$.ZZuowv[瞺QU-XRK7H1'؛uBpQ_`;S}{Jby>XXo{N5IM #Lf[x|ъzi[P>7/%o'%!w6'-TǡBu?otICcS!"zn. Uo &a> QT^1_F#n{n|(b1}u}#RR[oPf[ioe9ZZ8߹N1r⾽N #.kO^'t${5q=cxT+SK=,܉ nXֈgM^0[TP]7^ou/UҪJ>ca% .-K)$³DӢOu~rP  Xgtd4eTuxg 1ƔLg[*]YZ႕FVxg Ҩcσ^C|H>1?gP'Nq\*X%]J#r (Iڱh-Qqf#}U [L^1&ǫ"wj Q-/#y,|,f9v_CD>?;?.8{ 9d| _&8O[psd-׷gL0c5o2|$/gDQlMev.a6x:S:S~%Uc413 b 3&q]N+#9ν]&:aURxt1r1}wÜVTfl]c)xtK]< |dS=CԙK/;i,w \6:Lј#SF_'uOb ۥ|,$Ffa鰬*45NkS\{ w)&1+=3T;! Zڮ aԌeh艑Њ*A(/cJJIbNa@?F2VXm3@#i$[4 f-o)'pT___sώç{[X;qC)B [1UVȂr ƍ*JpN IEK؎ײVp=Y%+|h] ;W!禴X+tL*,b q8a}_NRd2ݍo s֦  k1!%lb*WQ`8=QL. 5q.1Q{{%V *aDT\2F^l%Ӷ;1aJu# 16?(3n[N r>k/-G=ܯ8 S:x̟ %iMD)=)#F*TX!T坃'Cc ,n%\uA )\'p:JΌ%;Kʪf-qRdEQrdt8EX Ad_Ǡ8ThU(fd*\oXx}(*xT#ƣ'%LMlN,⡇h60ظfsi\~J|"z:!c/V38gD`j8&hyxEl5wϫe`a$y^|8bJ3w,xT'cኝ1!\72SV yϭhI$3.Fvc5Oi4T/C!H'̕[YaAC,)#V!!ΝW)5uDLRM\e%0M0=ql|'@@–c'Pu:'.~!Q;&=$^iK0`& Պ;QUT4&.1z浃bYkC ̣"V@8(b c=+$ &!6M@8#DGOK7UK4,fl # 4fI*r!""m NS7QUf)Y9^"%L)Z?zvҠ;5stFҖ9֞lI3Hy]1 -٘%H}:{kEkݵ^[זpVcau(f:~k`:Df#Fx (NחB3*Qae#oAnVϕP21`*hƥ 4lt\ FEeQ;/n^o؂1.YjE :ʇ3NI8Oއ1Yb"~STLAs8;)yJ4$ϐq|tz*ڍmK"pӼ }kNCpҲoN !6tJ B AFM:%N .ɠ :$O^ftpvftqn0Hp$)}&ԒV*P,ANYM# i,[}dy6$䕋hLi>Jm%v+AD X}Z#ڐW.E2EBA GtJh,*ޛv+F4W!!\DKdJ$%bc7J= D&U[C4)?ԄaFt^:Kc 暲wG0T?H! GUʎ[{OY ;oj")YaԆ?FVm K"j~wLrHn&NZ?8IpϿw _٩}Mᬝ7 yC>g6ڳԺ3S"~9χ1 M1UhʴuSm*feGK>! C/v p|@FLD\buP3QQh޾ŰE~qqALP ZX>HY AxK!Q LQV!fp9k5C;%3@!/] au}}N{ZhK LZ25Xcx7QLj*Lr\:Njq?e(QM׵~zoVO *Xjp; R&4wT^6})x s~9"bz=W|ݻI<_/ֶᛛӋO&]sf_}rMoo...FLHnQ{{r~St2R~3 1oɸkLXm)NdFSB?.,ۮAgphqQ($NdTʵ\<oߗϿأM`Y5Knͅ9Yo]Y@u>I>,{1Esq3*5 pGnO}~25lQJɖ5ݦJ.dJ(lXrp\܋)#hX +T_mZ dm2tRX=7lሶɖr vEJǙ!f`|p56'׋etXIGa%UGrh߆eQR i۰,ZY*W&Ri \mҊO}p͕BI(lj.]tpe7 Jœ)_I.&h1=a~O3xoUp';2F5M MS=G"šH5rҤ9geWcŜwg6g|^UM63F(G!ԝwW 6/q_W7CR ^jzam x$R\65I,b)+L/HN4GB*eK-=+=ql)ill`@(5*x+`rV7b9wsw>"pMzlg[~?}AАIíU5՘mF4!MR)}$G &JyDFAB܍Wδ5ָ#I5%)vObjYpk5Z-2L24-$◫?~&=Gjq(@q!pߗ}&na 0ׯGnihЁ`ȸ)(2cQ "CH'-jȘZDgO&:_)>b,nj U Bp+` Wf%h!ϺJBrt'@:^i4DFdOB U{;\<}Ӣ:N8-ꢪOcCKdQ(B[k5L0j֛aPY,NJec= " iD%@5} $g/ֻOC NjEr 1H$4GaQe +O KŃu1юi&FjlC"S2TL"I00+,QibxrSpMFtToƹ5uI5 NedW25@e8]ߏӍm?"(vy鉿Y%=OxkIp;!1n>}D~S^n_}wz+I!^":T b͑onN/NvCgZ?q{{r~S|FAp53szN8T›əFRjB5ZK|L ec85U/O W\vǘ[-0?_7ɫ_һ%O00<^+y?8R!y8 ;% D4m y #r[<$"bLKD -iwɩV/)xj6eϹ Eb0P鄙(y`v^ 0(sQ0J=5!~bR?&  PZSVC9 q5$ Jz!Ű(B;2SS sSt& 5XM} 4҆27Lj NU6Wð(Q%4T)"2dF+#.#BƜ:H,lHbP'xO,8X8*DW[u.O 8DhDv{b,!je'R:V4ioNo g{bjaK1i*SlW MNh6&޵5lP}AT}xS[T\3;̺ٗj#S IY3I/C\H{eWɖ<n 47s)ܰՎFvL=!`L `I /ax)(=b4„p.TcRXփ7P<܋o[=`&/ H»X+Ac j`Vvvwo.w0nn1:+fpӀh> 1>MTD"ޡ95rG ΁aRAr3!gp<7 J9B !erPXkp:nG墍X@/M 1]4d)=&iןku1@9P{P^X$.>5S; WZnB?.'~ߗ$do _3ٱ1l?j?LoVҨ`TR^W~ V`d "Xe1Ȩ` X׷ssq[>fғ:JQ$$8҄9TM€`y%heΙά3xeZg͛q?3ZFO5Q?En1b[-[90᧯\jw$ab_9AhfR&7xͿo-2 SwPZem7˒kng6[ϣRlЫ5K;8͒䂑0;4|@#ŋy =6p䣵cheÿX?s!|<| |g j>_wy9C.Yd?^~F2w:ae3]CL1{Ph?L_l#Qͨ93>-<{X2j>@Rle mkzK{[ %[x8{s?)/v5k}nrg WG.db3w&ag6TYxў'Xl7 4:~)9Qo oEFT\dMNWmhf:ބFK9Mp_[sapN'8P k> o0 0hOWX_"L)oh"I9/%uF,URyrFHhl4Z5.k&PT4bI'{$a%6 ?|躌x4G|{iC>p9so\?/}^x1͟W_ȼ:vt3끣s4,'o..]_}(Zl/R7^XMF+*53C PKT׻'.?U7~œ)qȕHgHwڪ`qV\>-aGh緙:{#Xq-w^Һtrqfbq{uyIA!^iet_/pjߎ6G\=,?|k:Dƃ2`10"I|[t21/NځkA7 Dr xMrc Å) Ħh$ANp k"ETqcԍ/L^ڻM ;Cg3BJJ|7ї}MTo^P4juWh'&-`7Zv_2?OYhU3NsҶH췺rxÒӳ!8J?RΈLH/龸 znk73ak%~ _?C?d}vgVDKLnnk ]E0ru{/=|J AW "c# _&q31"ҒlIVG]^?~4d/]L:W1RLa :k0_Qh%2w0OeSuJ|6ć@kVfxoCRNTBΫ@W' \I%F>6%G{QS_*ms[Duz=L!QKv9+hvF8wPf{GC wNxF˖u{E+2:9L΋,MJiQ˷B} sϥR}X5r֣2Y h2K2߾-wmc9-g! 2ezIYԄ`#tXx`5 :D!kM$j+bCWQ:rJĒ'DTW{ƌs,eT#]f)S^N|)F.rj9֌'x.aZVKO4Pm#*|ҞNJjH[LH$$(kqٲÁe/j] uu A6 u0JE*ЕY4ݴ`M-#=n鬤SPj;H`Ԛ^dA[0!95oH".{ޔXϠ>!ԏ9b)"8ڃ3/~`D>KNaT쪐4ҕAz};,h6$K^wLʲābJ8S9*dJBmC0,_1֋ronGdI.]JcMCQKKCɻHhTZ"@1oO+đp'"46G7R1JG tE̻U RVY$~ !B˥7tzɀh A!0 Ydmb\ n-mnz\P9.~<Źu.li\ҋWdtKo$QIѴ\I.]f1bJ^s%OؙKGL!T|w,kinݲ2 ,0v>aoH) x.F4D #\&x @rg9Sa㊚FT[Od1cqEhQr@d|:)[PX# u4UIt #F FzhRV^5ނj~R#I!g"ƨ`,@@{BI)P(F<%%jA5Z J8$ %g$F6rc CX 󱳑 n/ .EaIjA?MpIAqtӓ%{ռp=_E1j40M@ָvWٛl& eν[C(yڷ5l5ck/5cӅ-œJ驚a_7׌۾aU[lF}3A1`8``*3LxcT) *~7aVT#[7QMIYrܘT:lAn?=ԾݦT-%Avr]M]~:SPmjSZa ]IKz D sf ΢ҽ1JYʷfsq:[ɥT} +)lt.fRHЬ^#H/A\1r:W%# BHslZ~PҌ UHV[%1.!)SZۀ/TPD8xX E&ƒs2nĸZ#k!ns`‡%L%SaHDgZpl=+e8q4Ն<ÑZrfjWVη-wCשJyܧ_C=>R [Gug?R4wa?z+ыƇ+Kֳ袞DsB:gt4goLot6vOE୹kŜ6)ܨ١,Dk@0GLgG($Xpw}KOFplŵH+1,.SٗK**(\ ԬiMJ {of n(\.@(hAєI5F+)[TzFGkW:(N8eDr\-56Yaݢg9.Ax-lZdE^ {*Dp+V5 >β7Z H&*%:nhlW5ٓo5-G^ w9 O7߂|Iiz2^/G_zߚLg+fIFTu?QFUIx?S|bi+ Q?ԫ4TvGf↞B{Ԉ5dć6M6HGJj ~n ofG 6z~I-Ms=Ai4C`NWH\ܗoDW` '̈́j;2IKt-TZ=̰> JNy |p, $shpJSDSyN n5Hn9YZ愅I^R.-r+8BS²]Rϛ+߱SZJY$mU|׉L4s2h^gQY$БHTF[HE ULIR&5I,W[NU_gKh@B|9#"ən^)8fR(8-֛Pi~Ǫ xy炏u-?r-eOfeB-+TT51H?Yhol")eeQT`e [*8Uhe-$s Ai^YSJII/1ժTè"zW%o,ׇ aVGV2U&aaYEJWAcgsAl @P$W#PjL2} 3r"p VQqŬzJ!&KbPz"0GJ^& " : ǘehP(4B$dQ],+:٨@t ^$!ȹz=VN(2PI& Ɉd fRA.K%kۭS4fȌV6V?@BE\p[ݗc6w{Lx|z]I ?<,1c'=BApEZ, <=/0=DtX^+X4o/RâDĀ4$o0 bpkf?xQ1aGK"b~iv#n SWc@+Ԛ8%qPHb&8 YH|:r-Md g F YX*s` bpĂ.+4(p!^/rYXdhyyą(D8rs$@P)p7:xE Su?^{3E݆'˚'PyMp^yE/j]:ݳuJ)CގDi@7mZ~Qmu8dW?zGiFSz6{Z^GZ:ޑ՘; $Nh28; tsұL"\gUDђ?b&hiRnIi/ɒhYh8/gb6tRx`8t7 ,cWTTEJ%U:bJ. >@<)&ѡ8 \`\gK9*+M oڍ<{<3UbnbR Ճg^[5J*Yʢ>nW]Ch6}4wPdz%v՞E~>LZ9>i[zc vn<Ѹ ^,[I"鯳wfL --hMߖ/g.`Q{LM~oP䇇y.pܒ KXW5^"]o.ߍ؇G[`oGsO&Kg["2elp[m~sϣoRu򭴕霳RYSr:~>>,!h~&A\7WSM4.|#>404F`dbk.#m }JcEMG1Er\w ̣<5<='yj=mq}n< ôys~X)\H^9b=ua䙗=ۥMP=ziQi*/CweAdO7̗_ۻYX3,`BY t7tcKH1ԺNnV:7?9gTeP!iWyvv<=e͡W!9rMxD0_>r f Gx@%C0z,3 ^V(ҫ)K~/xY!vGTJli9;F*7m(yΥ,J">F("&ҁ͟\£du,CU2TGɆQKD,GNLAU A WqvpAZEAdvhǦrDȆ@z$r=cAޱ:8iz4|01#fWg,dRuH+112Q-߉9Z^eZ!,u[6AIc:U8kĹ䎑V tHތUގVP6Z@WzFpD 58҄6['N,'$$@jRNGmsG}R.%q89cU-uN=[E. ػ*`({ˉ *_2bu8ۃh{GiXf`$j+"8g{֚=a{?K^ioſ9=.T[M+$3ƲM/~ۄmocϳe7욶nSwR N)oEmAtaH197+A :F|pB{.jQ :u`Q,UT#`oY T%}f1LgAAt닫fz^AgnaL7C۸xdXp+>4̰!q]x0ZP7' RҵNe_,t- e0o L܀Maa/wU~+TZpR[^\X"T cA*0y.-Lk+m42AxH^Gbʴ^,r].~rO+U/ޙ31Y8s0z媰:/-N8Ϻ)m]`YFrE,<,).W:B:&Xu2 ^qR&h6n&Z0x.@ QI]A:Z'-t>X+MP,UHmt۾隌dtaBdF*iPk0 7k/: Ba3;)^/R vƛowJRcI+P%9`!\>BW(q ގDi@9ܼJwXeX IP`z#[/XR-Ƭ+i! 0BTՅ+Q :>@$# C=A6c"pQREX >`#bqJ_@.bTsRPb1T3*ج5.lLDk@قXJ h,m  pYI8U)( cK 9c58BX:)MsR C yX8d)ha'W3p % 01Z{(Qa jf$8NI-s:X;xEX,H`AI} @n\bKd}T}UfeVE01z ?0k]i|蔗7ϗ3'4(S|PLJϽI~4~d)ܗbOq3Ag&v=!yI]6ҐtJңMV؍PbPEucv;\ 2eHZ*[UN5^e7!ح.UT;6a͟pAٟV?ZݚАttNɭ(|TQ؄7eZ%կnMh WZ:%ݯGnu1:u߱ .eJح~ETEvkBC^)>Ƿn|٭.UT;6amy Sح~cEvkBC^ҩcXe7Q̻ĠҤFwl4v(SBn h WZ: Vح.UT;6aց ^igحքpSk W؍'bPEucv;`[I٭ y*ZK=v5ȫQnu1:u߱ sȴÊքpSS[ Vum]{1?50pm t[ݤ%l!eUmeaV*;mm&:EOn_hbk% 5[mV>[Ռ񶾧iضpZlkAZF-AI|rL Ehں&-AOo2%5nsmQKzC5Qc~9f:f\96ܨ%0EO/ ѕjsmAKrr9fm17j ɓ1K$nsmQK\YBDscn0vO%96ܨ%m 1K)ksmQK'c1_BYFN/ 2!m17j z|s̒=V9旐cRrCm17j B1Ky17k sEcfl9cfL13scn$Acf6-1sɶ2lsmQK`\^9s+A]SpР{$72I| }>o.:gwf|&N^i{5' )Ix&RBR-Q? }FZUmow_p8J럾;ODѰ:>ԙSKi-|`84odƭʌz Lq.-e)' s˜`@ɝ 镅h@ ܃ @}~!3A̡1R%I4uX,dW[ŅӀ[X(JB ~#@ Pa.d&8azb.NB 1$D!ɨB*SYo3 +iM`!(V5!2s8Aes#(J)d!>2ͨ0ZCR;Y>@*uj3)lq=2Q`T5>p3}ͱtNJ!0Ճ tI4Tz ZfVG hk&4rj=*ԄfD,@0̩b --0r ᾷ*1R 4Cxu7OoYDDUߛOGp<)K&7.|##S &+䲏毞zUp6cNV%跠? a:8wgHoHM%r#+wjJZ)H9Ru]I95|".189*8"f\=+c%3LAH5Rz*NBYp4f;-{{7Ig^tV6tV 5)ADC }o5A˥Itt7/W?}1ɄM& ۰Q7?BƦa9 ߿&/.xů~E "RDC U' $i\_t[΋?{fOTg8`m. .iۋە`IAͯ՟ygrԸp$ّiz4EE`NǤ7)΅Anfv8;v1ZL*̑\r(zʺ҉K9py NnF-YORSA9A9\[86@/\`bX✇t-;ktIM~qF/2L&.}xX3HǴF7D5m8knS$92g 8*`xDh8Fv |PQ< w;98 PJ:tc4[p:> `G ec7!@_vUחz}o> (ᙿ_RLkd2rP\RI} OyFe(wi|ԄhϗYU^{O쑭`W[J\ 4kaOfS'Q҂٘\!fߏKІ%잤'&c^˯#ߘvQ;+w1#DR^q7IMjT*nbXsF1V͹H #jR)w_Ɋ nR$#>у; Dz78MY:'{ ̦d!BXڸY{e=n%;$|S߽' ܓpS黼k_I8OE ӋzGh,\igX]EscHyhm*._ @u }eo&H3YR6/l"|{qp U-h?3["5z;nf $dNLU]*JI6D$SYkUr-Ϡtè(T8%Ў!CڱL>Ɋ] GO&DiUQ Q 0T ы BgwYƥ12+QFPZ*\UD\IC`^+RZgcyǰrJeTh :V ro&-12hIqd4b3fJ/T0Ehkv3[RVqqCNZrO`q:l& :ʨ{eNij7 !@B#e7P~MrҞ鍳HK6#f j#kӛA_|{eKci4v\^=.mzifPteiܯ*f҈jtIA4ޭ5s(QADaƛդnK&.p0v%8z%Fi-"#X@8؛x< h N|[}|qJ _#3Z=47JMO؁F&&QHXb,!cE/x, EXDHָd8;䓫rƏc: u?&,7^?ҙ;ZmrCX.'f~l:g})x/(((܇Q0Qu*G ^>П +Β:n{HnA _fя#`p`˕ <V;y2 *ȩϟ5JnFxҡƃxK}oB4D4C[ JU甪kU QDl樔]c0wATA/DSX"Pl3Isf&Bny%RP`@"途3㴄f 4$D VpL:s~R Z,Vs c6R`VW[$v?#$!T=S!riأ ~~<{ố3 fiT jM~ mz>|??+k DTv$&4Iz"\ߌI %HRY:?l}U>М8"FYxч^zzֿ5 |&8gNκc5>2JÓ]|w[.o TsR*hi@@oH"#{90gJz3c&ty)wgbJ T<7V/"!PU,^$)&IѺ۾sDG6|@2Zw9ٯ.P?w=\{ (~⨀!O8$rvF)bUO(dIq$YyVZqHFv3A9f>־;0֕qMڧbplR lZqk ~` ڎ ,S?1,i +fiu9e^xNpLΫ[A)Wo߿0N&-ޛFS%K$=a׀_:PU_"r S࿀P*EWǂ!@t3f+iFO 0oul658=c7wYt/ 8nيEWTdqן&uY[CyfE=6'gZ9:,eT)SSVhjJx}Y8K*[u)iǏVNˉóȎ׀y$;U2EKAޏb#;+mGmVD钋[Z8#◫KeyTqXUrws|*!~?O{$Mqx3u3kE~V_ o#!^̄N,+VeAʢ˩ǖ8'D(EBENɹQt&E*Yz'R[CԖg7@F>#M )Ćjܿbk ׺<҄AHLA)k!j(4顒,iMcTőLHX}\#"M)ŠF!W.N7~RZ_ZF ۶.K{s]ۜ>V}mEP{[~hIm~r׆aF}"1k=jlovt=9=HDsL>ݕ;(*8Ū$*l>m+]IڲƪCmki=i$: ᦷk`oƻ5 2b-rp+ݱOpi0]A2i=M4'&z^J-FՊֽ~ePZ'ߊim8UF3Y~o؅ }Tl{=LQM/I6UچȞIe*ФDߗ;#voA%w^%tl4Mشي-k#ؓ.&d̞EuLW.A}9ؑQNɘդVZ&}ӗƉJS}`dnߌ XŠ8}nr}t /<#|O[D}~5jd&愐DZdϳֽ'mC#a_PǨZo:&p@fƉwRy%Tb8Mڴi9F`6]لq So=zs>_y{Ѓ?'#V;m r5#6ϟCT1cgm5huZ``+ɩRO #,0<|a˅M0ll `G̻,UQc7l%<{X3] jatL4kij{@Ȼw6N񝲱3bp!8jH 2:L4 _*z: ( 9v2zm?J Z*j:?`(-q}f|(uЅ!J2g+IcfvVm/;(pQx>-j1OENhA90EnCć4(4!1$$$}1M~VP^82Y!ʌzk'7*uư9Sy6ȣ_$l$ngr^· og*=L?yK]җ=A+4Bs澛Mt{0j9J]%MPe]}aq=Nx Bf eeq8f-;  bKՈo<*XD-k(zWz!+y cFQc 9G?;n{]pVK`&Iҭ$1P6F&Ű[,U&9u9W{CPj%._t0z :bcM6`D2XJ6"hG~ڑ5_rϤ0Ag>xroqnf=HP!O0F1i.i~jm䔟:eL$SB b2 AT!јq\V;$-_8pqt{}Pw- ZS*iڝPE:JJNHPxã焁j:% 6DmC]k46<ZD&I4N5˂ZP:J dP1>1q .Mz~=q1j&Q yArU}S aR2V7l  7PB#kAh9Ė8r, V^gȐ>hwa'כ͸S2Mklr|*?%\nZ? YZk3p=XcH9Lt!smSVVuVoqD%v&x8tkRYOƹ&QA ]t;n7Z*ݺ9y媱 )zKRaj7il/tTɸu(3BJ%Xt$ghUP,QL#FdeNZ{ +YWwnT9 ]\8䏳uZddyJlCUkbt4&\$enH"1H5A\S @6vv=wV+PP%.8_Y6%B}N d95Ot_\^~qYiЏ[az%Gi2!Y].Va#>}8pWBy-%gC嘃 J*'1iͭ\}*J؛oLhd4&LhtC)Ƚ!b:f(x9'KJ5MWd$AVIAq4A GTb$hhX;xCk^Ws!172ha0?\?AWlh/?L^Ɋbzwm6kYoO[Mݕ/*}|:MF:xvT_}mTU' {Qn-}nWT51**وs+y6[浠@W͡?)PBLPEp[= AJPqW~Iùi?ShpؤLt^꽢>q4>B}5i0T0BVb*SCf 1z-# <)!{/TNXH DVdK1 :g+oUnzmK G$s5tsF~-|"ojV;ƭSZ/^':!iSn6wqybZ/mDfI` @IiT64(ҕJTR2gMJoz7&Sĭx흸YhuL+τ]' $\_7f\B-kILzfwYg|XJr"zY(+,wGgAI 3ZS7ĢqS8`\0FO]V?{ȍ Cىo@}YO4φO XH-dqDT_J*R-ڇ_&|XoIt+%bpti7avABυciJA}HRDpƆ j&pZ{. D2jMpOްmYf w PiO@+c 3JHGMyYzu/4RFQ{cj-CJ3l:@p RnKЙ!LaGQܑr^pD ɍJKbB %`9[Gnjތ-f(5lgN<'^4X'ۏe?U>̞*wRV&X=9$&t[h1\s,!OdEȋ9suy "08?/xWϛn,9I= c"_J׮K9G{SWS5ZR{|k8aȅQdb`F䞯!U:d֗CYS`< ّE$`+51&xTmknk}b۱opw G"AmbpYj{<{ 7|hģoBJ+rU0ۣSs2z*Ѝ.'t4ҽ_>VM8IY\(n9::r,Oؠ( }Ve$Ұ0#= O:NJa3^Gay$T`[ 1ZLj^)|\JIK/#=A`ԛ(s҆HNk-9LҟG=&Zg6HjbPU h0=w bE8U9&2^D4B ;dՊS Ι'U );- d]4q tnwEf>8䍳yJr~(MjrVWJ316\ zåu/c\@2HxdudsB,k)%2G ;BB)Bp̃g 4I^N<6ˇo3b7bۅHJ5^0EGv0ۣp2 @Q O%Qu͇ӫu9uwsTgk=5AUڣnɬ4dw,tCu+xwES{xO)WL STK~Dժy%^J/ۮ"o6 뵒}QBprνHKCVH1XdF+¿re=kuto Ï@@+JX$xx[ s MZъjExǺÄR0Hm>})P ds+\l1f:%elD3|}xf/_/QS_f܌T }V 5W܁VvõIM5pFdC0_3!UuJGb BQƿ?ICR5DUxiAeUmZ_c.2Ъ"|+[̓SAoL9DmSٮh}#Mq !e #8؁I)) Dʳ<vDjFzntqpL\flUY-FV/#$B l퇫6"]RjL-|G=;pzŶ34zᷢ%ovbF%& 5Tɿ%eU,V\r[FP^q|3yq(e"}YU>r8i?h2ïxpHϳ|M/J8ߚ`̲l 8B)D12_|eO6J? Du(9'"[ &DN{B)I'׫,?ާRJ%LTH;o\E'fW3_y5Wt o)q`RFy+jJm-r*Ło}JUu.YtSOb犒CML| 3UOlY9:!v)Cl'jvUҵ}Jx9jAI <8$h&S /CumhEfqS5+sB.F"=21P#uJRjqZ)Z N7@;΅?_L]\|~bV +>zBPkd,pQQ[w ?.!$ywZ7h偣7tK!Ex[ʧoW5*ekAJ5 4.8FCzwtJ}Pj{/tNy3ķz7w=ڻwUm7_?@j6_>ޢroXDP_zr/)FQ*\Ln|.p55B!5ަgDU)C pd$RG>{5TG伀F0KFJ%JAnLՓz"s#ÉsÌiŒ\0#APJf05{FTμFvߘL2Ev YLޔh)D+tMюvI!o?T*(^jYƯӖ-mHV9>?-0d?Kg(Ïr'f[+> UQq*㬬=aQXMu`moi4-oh#xԸojI˖eig@H$ o7a3iZ~pC4U#Y7ʘ«35(3Q,(Is*<_# JJ6ptB@Ide[ [ S¨pd<L(Sڪ~X+e)iNGMT^(1\E55}O;C0RlI3 %BMzQSej7Z;6*HƭIՉi2$(H3QSp_uLEʼnȃ^S&Ԗ p=Jڢ&љ(j áWL@YrVt19N_lC($@LfY/=~Mrs'ԕzGAda[yNp-Gk!/%sF/W\tFS? ;i׌oЊi٫Ȅe c6a9I E5\XT?,&JB̃KbɹGLg?ճ|OJp/IL WaSe̱l$g-= AC֑N!Eeڣᴲu Zoo9st3_JFAEOĺKaؐ/ocnib9J֐~zwCzGit |PRR088${fzRq7uS40DQJ>3_a RrֳCrP\*G >kR\-SOೡ?vX;?zfrVl4J7Q߁V"͖2)jm|&yC?y2^r~tmՂC*Z?#]}U,Y~% z9IlJN'{ʬn/nJClU)ō3iݪtV%@H띴{nۗ=p=Dm4M`QcFfjBލ@^h#/vM86y3FFO#^bAr׏e4vX}{vWzwd `y=}S͙7C|; G<.a|ѝw17bECMW B%6y1wf-OE5B(z1o"U0tN( Ov᯿s 1NRߺhOI!^}y"BU;X Dɴj\ODN[f |t%_¤!7id}O7xIG ?5D\ g7 E*Vʌ-NEI:zSFS)QvC:ϗG#,=˵Nt9i6ӻG<$4!i9Y\֤nדx-,8<,"qTVۀ[`XEvS51A r(32{l6{~-ӓڕ`,Lr5!e)~)d :k jVsηK?mE9%}zv3c}~e>\99c[3XJ.' f){a<^”q:V[Ue<~ƕxŧòTfgG՟t{wgS{tTz4h/Osg\gg S +`^}uE`o3NO64Lh Dw7B⯷x=߮bXe=<*e%WEiz귻x~{zv1hϚ[zBZ\BJn{f2=*JT[ac,WۻW +'@9$ؙ8uuwt:8W )928K5tg{9jW7󴈧|.J_9;c>Σ::9C20p^&h;1%t: J;B-+=wgj0D.љ+qAt C^n%3b_drrf˭ċP-=1^G.E\(ϴgQpI;'`^3B|(7ИOxO`w\4k85r9!o,HJ=d ,(,pm+jT@w٩ Z33|lSW-DlhZof+Ƭb9!Ui-SIv8Ҫ&ehZAզ׽BEZl7S LH<Uծ痧.knWlB\n;5d.¼x x/g:GM ;͛‚S/rf5mw4g?C[7eiDP2a]SknwRymYkO/mD9ߌҐo\Ekt :u'Yn< VA~u;wZƶnՌZ!4W*=5fbP:cy ilVhuBCqYr|j˹ytY{n.>CrF,+7䳻>9?'O8Hmۻ3w*| ĭN@D1ι^w3 ^9aqcRJ c,č}rbj_nWCzQX%T0-3G4 0t J0'&kFDHtB ׏L/3Э SIv)2=duxy}s l\/ka,t^dL>N%<,T3p]?}>Ok6~o䤸ӋOm:X ?(/rvY%NE; "!JHRk4ygB arE? 6HA0-tΕMvR*ؐ4IjʏO\朊N' Bp];/ x7 UV# oӓ//g[D[="`\id0߹Oˤa a8yhm v&k7L+К"IO!j&`r8(e}1Hz6qY  0S@Ly3.0!," ^Ђt+C@y`L88a_B@o}BB~KcFDLls@ w^PkR>44qf' Ӳs~^TSȆr38t i-GPE{@b): -I9{ȵT򴦵Oj$}ZhL  mmR mcl5\B jE20\&gΔj|%ubқ7UZ!FejkX.60eqmm-P}*l5j '†PfWoNN2cКs6QsΏfh6!ù7VKC.?րlM'Џ-{w M*7Bh_y q""Ƒ[$"a k/߼ xs{>祑d!eY@aOU`SlU>)6!N{vLCdO/s>vgCYa^O/ȃL)4NzVMX ?H'$- t-xb4zܯ44rOW&#+Ff'h4SR7K W{*tfz"❎~:!臬:ڝ]w֓_US0T2btk94γ$8Dpd 0%dBXi;ڰE7mD5֏ D]æ|v▀|ˁJxHр~F2p|Պ{R74:B';v=Nv%d,|ix6ܨ kTV-TUj34c*|M@/-8JZ>@f7c b0.]f!ULJN R'WOsU'Ri^_ .VO<,ػvEJU{ x.o +xklH' xrF5ƭ#z]r'0_=P^e?٩eKO~Q+9_wIW2:`|"R+dRH>JnV-O[%룤xF' 3A"JO*y)J0,oV%?7҂(m;tNypsM& [Vi &,+F?j]L ~=ͺ8h}%o\'.µTrT $r̞$ p=R(cn|<M&!LɖoܹWOJ>? lOW&#+SV=U6;t~H΢6mYn6K#j}^Ktk 7L58zG]ʹ\;o5Oqo`Ϗ A*ŠN@+c'f(KicL#}4RR"8dQK'I%lU@s՞0˃_U5^3T!Z2稠a( 6ȋPt6Q6b@3Z5⫪& qS'רZ_V5ۡ(r!$F ABH[-O)CFQ`j6$A|DAms = ;&ry0%hgsQMmeM5lQ a =|}&UUBapHa sm.d [זˎO[vk wm?Rs,=\ڵeN ][zITJe$IQvCBTuԢg ME#ئ~ն/za1/zx1_|1}[#[>]zo-|ݭ`6z{3$gaۿܻگDXmFF5 ^AM|&bE[bۃ`|ang ./Cfzݙ3L0Fփ ùӲz %殊9ėVܪ)sp>FDXKC0*%'e.ePwW" yM8rbxLp!.VՈ)[|;Vw^  (_ 4T‹s.ڪnTh >mll)dJazead=5&@![̃+y$S[߱}н|eg2ӜPm6>k}@D{ҽ|{x6 m߫a̧Cgi:\)Uf˜87$'(*cQJ:TUl׾r4l8jR@Dѹ%+p`{ВHA0 3WjO~95UA9%5ZD)'mg$hN[{#QÎW1},]b>Q,Ѧe碷ǷkL݄9ޞ-h3Ҁ;Gi?ZfJdi)?7vY`O s-_}4CitbmJi?e3s>M߅$Ӵpo~iF 椙6onMGc1)%ccC#Ʊxc',3#'4>!$ivn7)M(Hm|ͣx|wȣvp]9I>XH.f6ucDFY*!rqAiߺrz*92i D80GWl+5l}fz*]?ُ2֭2wF; E]ψ?Kt y}ǜ,LmqM,Kj11q=I`&*}}8%HӐd{i "i,(Uvni%nAgg!<'ij\z64%+pm+fB+AIUP ‰!?C䧀MהAtPIԲȣi |Ib df>&AҺ#~gNb]F_,ec8y]_1)Y9Ň+hy[U~(〤&!-ߴZp*GxJvv,hr ^]|xMD֖"Tb% FבּB8g{s 4H彈ZFj:̴fŻ֛ M,u<:ʀf& &‡cMk#ELATw=C]8V͡h'97? q٫F٭`vRѷR[2*b$"2@3n )[;+9NV1*%,Tqʇil9"J2dA0+#1w#7g{'kxө?;lyr&(M˭I"HXTMw/8g{0T"ڞ;<˃F85|@HF{`Y!{ؐG%EgPҹ MV𺟞]`)VS?6 .Jԅz %p]~ycԨt StADp+J}Ц'h'vy#8Wx1R|'BڀD`H6YZ5Mծ",cJV#VQXV2(fQ` c9:svnjGHmA^AQe~ JYāsfzŵ!&NHI%'z""&*XrЌ`qAYNjJnεzl!=WvFW85Rh,o:dP4 *^_B+}ӾƑ=W%Ѐ̏ k s+c8V") q;k丳oւ8g{ŸbC୒Ip4-vJ $\>?M/( \wqȥbMJ" Ȅ&ѲVd2:ׄ' pR+x.?gquGx-\iTJUĂWҫt;i(T+5rXT`"Ѧx2{՘f5۷P{~sw I>ŘB( I)xЀ˯zPf*zzQ78IYc+逢^^:(9Es$I҅)d,9rgʼ7$rW_;  Pd1)=Y zg1Dml 5mDl:PKO2I; kl!cYCNJ~Ěs*<"!C;[dY *o{@ Č>Ɣ__$AR(2rHl:X|VZLARNjЁ3vY am>\hrE)c98^r(XCU1ti% V2Z38 tc<Snj@^(zbׅ6JTc낱uBb왭8ȋ8r Ŝe$L[_X!xf#/v!n̊ϊ%<1'.vx:]`R9'4KJ5[~5>g%`:'@8x2ju7vRq'@ ݝX6462 :Ot 8u=7.y.!J@~).W؏b.ROtVѩ"d~K̗w_ hG8DQ&(&A\)RzHbHt[[ǥRV)-륀9թ~$%ct_`LJiPZ}ƒI-|27T=;wӲ 6̻»Ьڃݵ>[ 0 '/h~En9(L {I5IhcՂYGQlv8L/=C ۩Sک!۰XomgڤJ/"7fqN+%uLKo?|;?/!ٞ?I>!ҽGHaD1yT(yĴRu+-qNZ t}}݀D2x/[:H#l20b#KB $G|{Y7 .$ˆc:"YLыq-}ddlaud?n]NYq'ǽBV/` 2R1tJZʄ˜1BF^ҶK(roߜM)@?_RU7uU"'PLKJ!Kά<ٰw͢3DwMU'nxIkK[N7 rvޘ`e!Dq fc81B # `o"8fφlۗbSj꯭ЅL6WyS&9 jI aYs֢wt2д|n4ojn7ᣯƌbVn[zEzQכ8 PӴJߟ !%l'PKTx$3f%g%mP=t·8e4ئh.42 lpL%y.pҁ }q%!Wy)lQ%?1O~ÊG+v.Ss_'+ ί$ !g$g9N!kxa 5 SFoFkڠwՐ7=Ԥ)S,:;p`g̝#מr=hZ](OY@5u dsދ a4"\4#y Z$ W.7Mү`50wDr*rQaѺ Q0V5Cl%ԡyxl!3BUx9帶u݊|u )%7[! :!vu5YWs:Kƞ%,KQ )瘐1B[C[:^DF&|A๔hpefhÝJiB%d(OxNMiޣ?m*RfaLniG;nYtht|<R{YŚUu'Zu`*YL1 K/D/zҢI!ЌCZcȅ)0]#KP!_]WH+S,+\Lf2 zHD?yօ\kx_y*D_ SnRpJc+Z/,jD0& *T$jE ^XJ 9.1TQ-x/m|R& Bvzvt &: "Nhd?| tɽ理OGߩ!R{{4OJC@ޠ#}Wƚ51SO#S1HOFuzjM1GVjgdHܣO| J@<#.g,t[lh[fX˄H3ohc HP\3]oh]s2 KUZ>RF˅VHfu0G[ iBbZ5ARU?,=6Jp=2:,o'p3SZ$θG$N$0fũ|yȧLd8C4#meXu;u;/kG^hYzwsT@b^ZV{hŴ3\L;Ŵ3\L 3J]QK|T(/yO6QyOM&Z=^:hfl$w Eׂǣ:\psvGYэ*1eR*qa9ѧ \L@)ȹ l0wՋH|yEךHly H<҇Ѱ'Bh^7:;:F`."Jl-5+f=f>#*a]Ld3hm#%s߶v{SO;)zwpW{oig/In \sf]{gw{7{]=W{q o~~o{{8ܓlwow݃m dؚ::B3;'w;х?=km\\@ط[qc'ΉuOJ&0Woy2?ŧ3HnC 3n7/sucMEj<[L Jo/ʯ/'lu[Y*.߷v|2;|>?< wlI=* Mr ~=DxT<~6OşoCo.TqooG&/}bLTպ}~?1hL+W[a晓W >}GE쿜<^gԱݝصy%- gn?]~!Ԇ{0óB o3y ^}Ķ^_f?Bcg|@ʶ;FVɵם}ߡ}pc.LߡnhϺ>$QWMT>c6ݝQo>xiOt|p^?]l=vzZ7{;Aϐ?0םٗbଷ=,ݛ~}jg_pǮtujkx~NA|aŻw O)\:nnו?u8x^O۷Px"W$Ty xŝ=ԕ[ȝtl.c ;_-Bp";'LN:bkH?")Íe)z }Ʃ !SQ$,x K x8yKiTBvoUp΂k-Tp >TG:\Tp Sե NݘJ#ʣ0i%Jm-.)jRmY X1炇=ᚭ)rvf!K,vߜp ߼F jE2׵ᘺT%7(sDIXr3պx1%Qd6D !Б:5~k !Dեly4*l>߼FUE_"G ,3u~-O,W=b Ԗ'I)]c=&ș3D_U}Ezňԣ0RR;ebhLp%Ue^x^d >BnE/>$7W ˑC,WwIr$ei,e>邖тpn8v#L)Vtp Z*J#A2\pkfÒïlmOq#J;`*U|Tw7,B3Cp_No K)S?2JX/]uE߿|ȹi INt&YE/g}Q>g}+a<KM]T,sW^\;&89o%QIATB|E|>hd!JUp\z@Y36$\+an1:F>1:1\qy䞡(2lt8bʈ!Uب $AKGdbӝdB( _&HOIb!Hͥ`|S82  lj@kg0UK xh N + -S*KWbnp]j9SKQ4"fB&:"%8p/>Ujq')) G+]F`K?;&XQ>NJ9Vϱ9JZ F2E!̰f*8c6qe@!Ka ϰ9 4Ņ3i+Z]VYzzmb%3YFEB*d=ձ_:jF8r#ǥ̀/g6kpD@ CKChҜ\MXx")X F C$)HdJ !K ٶNZYQjj׮#"Ƭڳ񬓹ƀ>K}x݉qg0Ra$NBAU 3p*ZV>q"N ;$\$"O,͹š:(]Zp|Ħw=y{4yu|T77fzV&Z{I( 7FF9Fګ脽75*91cqgdFHLYZṄqdI!RN*LT G9 B@zsf&yX$ k Q : Tu&f߭PΥpn(J(d-3cf ҄ CDÙ")(a*a-u(\pԉf 9;wĒ@\n$0-<)>((モ855 QIڈU.UgnFE&ry!yZHb.ƽRIrkC {}pEbɲ!Bˀi`dkT 0w&Z Z>~n?eXgaQuFeXJQ&r-*Ly%hi\C5P5*e֬Rvä"3 IH D0\> Ao;zy΃vy[Y%~$OGC(psc\@D[xi))du4Y*Ό"<|M+Bu,xprƱ6r,jZuzipG ĿgJ{J_3Kj_dtQiiEiېϩk/ ظԭ`_߭SyNEk[\3Qy=u?>Lc0F75I^Y8H c! ٩siݍĚ<,>.Q ݟDɇhԳw3O.­T0PYf?H<8%9{V ݔqNh>{6wU~Cv>'K]hJgy-^y3/%yey]c)pԈ!d1xŒ4&*Fd69wI#0\[#3t"|- ׏d Xpz\3`=*4>{`p.YħlwWԽэzЎbJKب".G ɤ.*ޗbKė;vgͫޡ@ȝs/"_ϡ~)#|:>\]*v~a9$OeJ8xhƼEzWDndcЇ~e,| ǣ0S9 = ڃX/^)zyhMa sFZ/S HPI_j3Fd_e'x¼Iƿ`V"z֓jS*ODJMi YBϿc󸿌/d,W@mPyEȽ%T:=Byc:"5iHOtLO$2iNL'} ӍЃ;l >d_2 /׳s(NAŀa( ]Q m|ҁpWfr|^F9>ovɀdu]~-y[)/DW/o5^85_0]nY}ѻ9wPCD 'S礢0Aw 'XY\ҧ~bZꃑ8†bUzx?\6]ߡW_Nrp^ِV@ggjN&iBzϣ0'aHxA:FP(37r8 <\5ۢ?,Y85CSmQZa(t_#mS,,cto،@*Q*!9E(sʂYz(شAI7AbhqF?u3׿R0`>P3e;Hjϕ٩h$F<( ^J╒"J#$EX"'Fɵk J[WlUH5瘘yGYz dj&Q\8SLm3vqUdgJyЮm89Ul%\PQ Zs#iF*$ۀiaTH m2?ڭQQa4RTpA7#$LN{n$J42:΄L`D(j-M:dQyZe2'1 ere=7kJTZZ0NC% ՠ)=mLyPgAqȚݏ*Ub. #ZZ^#$֋ w)kX+s5Gov6LzWp מ66^5ڱ! gr8Z`wum gpR\rbWl_?$.|)@v,@KP"lw>3d3YT69@[!ÂH' /ʳp]I_@>݇UCg?;)+d/ѧGm~D+~`D#>nD7"uhG_/E]I6Dk~U>>-ƳIsc0:ELr V.\吒1Q7)ԘJYFT,y%Z4NfM]yCx=*x2}ՠ#1[.///Fןsq2`_Asū4Q;WׯP*h0A(Llm BlV$tLx>gLs~C䭹NReD`}>fwVy-YnOΫ,Ʊ,uьp *j7b?J7=/"~tqOIbzJb4Fֳ"/)[ˋ֋YbSN./j\m-,*! AX*ůuZf.&mp^nd?i^}/^OFCl;+z+xe߲jNSv}xS379XԼe߰iXgPyOo|1՚{NDXeM|u=Ŝ\0*Yj'Xߗ?X{u[$>tIj7*GGFN;[-gbcl?U #.z;7JiVzhd;#7~ϖK1j-Rgp=e6YI!NGH'SIEy,e#L6(;%D>%W)9"KK,/.,___.N"-AЉ^ld%3fLEe&sgyeviO7Y<}`.EМN~\;R kC y8+MLjtGEtWzVFNYP(2:B}-OK2"(V ) FijN ]ƒDBtNӚ~E`5랞`m8)5ơ_4Vq@/r}Q d:+\hYQځUuuOmmeVhoóv=HeܥI-v7F*_bΝ8޶޶޶޶Y̬9"`8E`2ƤLJpesbJ<3b PÁ[KwֽjA:$ᆏUչ2`8kʥqsH/=4DT;>RVmQANEE!;a`g R0j |wU/)%¨ڥW<|P "Q$(tc Ye%˜c4\Hҳ1AYV!Ȉ^'k;1 I)ē*O!!e'͞ k;r.=ҳvFbMs0Z 5ӛC" M29DdLP%+*{hV(Dlmcb1攨G= 8e!DX ~&> z %/qcq^驩wMaɃjj\}0{H CIK)T&ᝩ'$&etYz-|,Đbxc9gY?wN<_ğOSAB{r%R-=J׋'ހkwJ4p@S>,ԟ ;A߯'מT^{Ry*]7JJ_ b|A9* [zmR2πȑ-z:Ѡ~#C'-"&ҹ@ПGs 7 'ˋI|DA)Hͩ;bֆҠbmC`#BI1Bܧ,c*"SBJ2QF;4f<"xɧ;蕳+5bҗ堕P}]AzRh֭`>=SLGY[u$kAHBvՅN׻Z -ljIXR;ՙi570ƴGymϬNqGf863%'QHBU W<5g +րJy82Bd/Gɥdbl]ʠBv oN Ɠ Gghc/rrIp` 1xA09`}3`b ר{CUurݖO%vW8W }Oa񼁔{=ecX4G ӣ-شbYqX*Ɋ 8kТ[7T}~MPX m଍݇YUoD H9lc izBivz^  ֜/͑zMV6A7Ah"{H{5 sE՜Ox2:rFΉ!PB,.HR7ˊ=8]zq0y|^(╕1 $$w4&Sc+GRM'_bQo\-,닯p gwqNvȮ[GeN5e_N'\>rDߧMYC Y\CRPcJMΙ|,D뢄Ctv<pZ9g A 0|?cvώ&Ol&Ia ՞=@*!P.GA4C9N߽oPF k5&43ъrE`{5Y0cc ' !qÞRq=epB2o4(4>J  dAq+q""u z#g aQi m T}i)s(6l`YCmʲz_;=z']C-S!fQ *bbdrá;J^2:(1ZIeŮ_$#ٿBٗ&k=}L$R&)94\)s*P=Yg{Fkd"~K۹JV,NdtضgSnS)6><`u:qr9HQ_ݿF܈1Ρ[c,QjB\NOgnph/ LʳɊK=<)$c4"wh&;2eY`Vv[#ʼnT/|k4bT[$fA  8HN.4+T}ZJ@\qġݢe4"7\O\I0 % iӳ=4RMN߯Kt {bݘd\<bf5 !`wm.i , za*Uĩ ԈS\)'QޚKCܧj0 B*+([/gVL9M~09O ,1~yl;}^ :CQÜVyVĊ@p}>fP¬sWxph"«L&f1DlG=RX?k55jr[>o6ZC#cCarr{,tB+-<x2/dWlf|X1j/՜(>n{GJl5='7iQCϿvZ7:]rco^Own8#UOKl5/,5o <_Tז9;۵^'Șs9r,-\pFqQAksks3p#=܌JpU]ZH:7F5+)[|*Ic_i ݑzZM?PBTZQAT .Fi씁}L95R1+7SPjQQ J^jIt_u+buo:ͩ2{iA[9qLHjV.ؙ\:+M/G]H˨#z6?- |oLf^j7]LlB]E>b#1#g4\!PWxjob6H/e#'`Z7)zJi>pD*^M}|Vx݈#JgD`3!ƌh0?#bhΉ$/XrNRƱh N(cn5K4y%-. 3)K-b|~o 癒O7n gT وVRi#:ϧVivͳ!PM5KP;w~ 8l]N~~|jJh;P<>:k"}$JL@$e$y jL s-59\l|<~jYt8ob4_\Lqn>8BתaɋgGWnqt6͏6>Bղ[Ggo#_OhrWQc=iS%!n-,Rgq+Ks4G>K4Β>w1X.ѥ'P]-aKK+L[2|.EWnR -c+bY.BpBw@–Gi/b:~S-y7z9%uqg7l߯|ߺQZGs6ώJ ŧ>fލӮ1Hr/Uw`a ط/@+X,]xi.YSn̉"z_\ʲb>ZM/-\bfP pBr%xs6:},y\Q8 %mTۦyԎGR˛eHx*lEuigmh\Ɔhu꣚ovkdܮTGE%w1xrkDj"6(LN*h2[w>&*g-jܭYFk2u,y/AQ6 @@s\Zu) lOS>lɹ-G&XK!kh"у]_y[ϖ67I^hO"֟ޔSRZ{!b F/"q@Tadn&ӀI !T,NK#"ꁪYfv`M: ( 7o<|d4[VgFF4j= YH?{EQW`]`=8M\CL1LDj2g(ْ.mTY.\¾0@јuzf`L¿J'9lj,ن0= a Q-}g PI>Ojmz0aYȆ:a4]l$횂| ˾S<# `u6zGb=`"U-~*N>v|@GfNg:NZ!D=&`2Ks^M]C7mB#}]uIQ۠w\ڒ]}g˫Z-a$G`v1\wpԹ82{h3ꬽI;ŵ֖ ^6[2Wjqf3"TMF1%DtL4ȭ" K rB`QH 3[ĉ:KЌb/Н2mOS_5U'apgl>nVcoFJ~>ۍfX}ϊ.78"~BAgs)qݹ'>xuy":O>]fu iQ>Kڟ{7Ўg _As!.Bhnwm)oOF_;0Txed׺QӦk ɨ"Ǜ_<.W1VktO?);b:TT$/}TF5w2c5&0&Y%H-:b^\w'u}m[h$`uX< o♣$no-BXsYczk0wKSDutevujiNՏ4珯m ڠA'hN6s2IrFTF9Gx-QeQ=5,R liymUATUׇoGRǒizW-wU@g JmiIN51p`mvںcZ b=CxlqRWBt3qW_|*8zJUg_<{>n9y?mzeSx T|w y_ŕqw&jyux/` vR<]EYC/Og"> r6M:\ /|7_[^GKY><~j%ESkϮev^me1-s=itq <H7§ߟ]niW_,$LfDʉiu&:(II0Q^'N&('L2j_~Tcޯ5ly kjn+}6xq]7K)=Ntn9Ƌoǵ3pBsSgj۸_a]֤ƻ|U^Z_rں:RXJ"MRIk̐Ԑ"% gHI'&g@תPڅ-> 6pvs5G//2rq92KK85}\\W'|o.>&YyD`mT.~xuQ~( 8/a [lml'/p@Au:-sэJX܂ߟƋ)۔PnSa }&ëMPV^LS!aRU(^Vkٟ.|~iǶg<;SϾ^ߋׯľEڇ+^^]fV}z1q<BH9ߎލFJ,pJgN?ԫ닋ѯW:{QUP=o(LesDѕ *i&]]' Ծ.wub_G7_/&fγ\߿Pr\/._YexBMrׯbT+;2fY]Oe?rL,=u@eQ-&)Y9+.(K\YSTqYoxM\Z\OJ-Tߗ~)U\w.?ih/Nd+I=LWP L{i Z0IONoDX,uPǖ8y4fT6ğfC;"&^|IB;=Et@_zѧ^ߟ^fn[mQʮ+pL瞯Ik4-@֤*QH|.ʻI-x~`CڍŚZ- .!' @crvl|X\Qm#rYLـPL)a/?8~^.o$Pp:}ǹ4/+2aDe[#33#j$Oً͒!肠,#"HV+L*@pC n{fhO(B`5]k xw.F2S(\Dr)@!"Z@ AZ`H!%)In4"Iuই*Ɨ%2F^Ke7 !`K}~ـn[ IƇMĉ Z>E$R2`VkeSkWF;y/Gn2NO[CUޖ]{NHχ}Mo<3Tl8(;=Q `R?;k!t\?]>:coCP`oaтjʹP{L̮LUWS敩V2U*4_^ݢ^(^J'(hM Ut,F~M^3u%rMW^hHg~%Խǣ Ok5}*ҪPM3}P"KJluG|FG#>D[)[MMnm1:mD=Jڞ[{ltbc@=瞴l{RT:/zPYfH"[gc|ogPj@ji dfS 1`*e:.'*󄍅?ɊNR6[OǠ R+x 6(^ePPS o>l4915KŒ뫪:aP.IAL1]-5x+GtV9H- H;)<"- Q.H"@'wd>w1V~+ZɅ; sgV;Ռήczp/IBVyI/8pIG? Q+iasOSPv[ lB jQN  B!s'ш"H4ɼ$hJ"ضP(縁"vjfvt:C`ȝ0*(pC8-Lf+L :"+^ Sn^ܠpu9WqbOlwI;NO`JV}N-:s)Y"#"opI++KTѯ-\fIoC(\ޕ^;Vt:p/=Y0bEJЎ54[/uRzG#{vNAHA:a@mvI0Nn#"cJc2}4Dk$*\0*ZY zyOgR.I%2qBkZAcu"2 HՁGEp`2x 9YQTQ֨m)`J\G,D28m! yfWhe+yXvKE! bѡ"1K4.BsP#RCӝG6t4 kd*V5F B=嶠yMG#n9@@uٸBkfݱ9O۪hLI[#ݜg3ցw5c$#xZC61'-42F;Je"$g^;h2e3mĠ@5hy;-kmk/u;xڋ F4@^#vǵ+}N bxfL"l>+qInΘJTԬ:Hɼs,B I-Y-AIDza8ػnYh]#Ρy2Pp3k Bx(g A.-^hxCiJ[81Qlo 4/LC,ykiwwȵ+;^+'^ƹ^銌.8j1v>diN2V5f jWt޵Ǽw0յFPNA- с󅵟|.<}逐{Ef #WP%LB@D˟9qQަXM42H)+~-:e`Q+AµHPXB M=ތ RNvZ-aǜ\m$ 4 Z"T"&c;JKlN?-N@툵!ZυVhQ\TZ-ZpN0 [^( *֨f2j& ض5 "YYl9ݚnJ_Q1e\pCf`ܨT.̘}g;w iFv<]#Ga=3s8}Np9}TRiOO\ٽ}c 5!ůמoƪyƐ}jCҰf(:f˭?uhR;|i9}Ţ-xڌb0}JW+s0{B41/r q߄ Ec08'lhk jna mIp!8 rFӭXHÈ{@"PkuВX XewH%'q^Ni{#ԁ-F?=o9Y .yaBqHA6 JI4pIQ"U_dbha桖-(-bRI"ze{k6~5kc( Mn #\ۈs/W2B! ĨLxn(8' PD ٵ%?LWOM.*Koa"*:ԁtd ,%J~~9uɜlL Sd|`P>"mykowr&J2oxգն"s e(] d, [ Zt5I[%_Tjn܊i-@JJ9wꡧp*J8*A3!\Ҫ{͇ynh9akUrn do0UB})&nZJ-SYy]$=/D3n!%+vAf2:I҈KVD5Igښ۸_Q\vt7nI:K틷\p-Gd@Rpđ1ΐ2D"5nt7_ .FGH4B9s k8,t#vhMϛLoQtܓǂShPFmi,P';э?e) a-v=VXԭohv(רE\uXbIٻ=?C޻|Dgy@o| g?9ʱҫ?ެ[CpW~w~9y=ȥqo7,ֿ/M[*@w㱧Ҙ\VõKt~# m-5XlsmXDɔ!_)!G7bXO9uʃ.mƸK1b8u^ XhА\EG[}m(JTg;XOI8urPhА\E:UX/p#@5?](P8638 f),WP8tK- F /B~U<b~XG/bX9g Ű9} EJD,"]M(1@5as@ʀE(8w&|ijgCy; v5 \)01ɓ vac _f6u}6Aq0,~2"-;0|Y6dCJs.N<1A 9L)]o'?şC |Ȟd^^w | ͯßշiA&/$/RujdaJM=hi\g=5U*2x>"kGlNA"`baEؽn#f@*OC3U)b%T :JEZ"̮,&OSv Gb$Qmh[LvR6Y A!g90V+!;x?S&3ms(3>/3mDK)aVI7K·NQtbbvaCW7hq ؝¥e ;ODP3lq<بǸ& N|Ik"-5CR2|:]4wԡ%JP*-GT{ɸXe̝b7H9WDݩCX8=7Ӄ57͝gvdٯ war:!`! ?S=*S$Y[kWxBPaQɖZE$,*xwYG+RCgAupp Є4o^/B XO^БX2GiD[ґh%,Ô7̠ZWtΆR XR( ע_voY.y]l`aZ,~2Ҋ18pY6dC)!=W;^H0O+0Y&LFNP FPF0J'L%-x%+x/t"J{C&kP&3!o8F)(F3BysJâ1H/i,PUiP[UEQUYҺ + 4iR3)6M=B&!ob ˗"0w;LV1-]oaeX\ 4n0\Z!c2N cҥ 촮Ym=3z҅&]Z1xuCrxiR(Ӷɞ(ЋѢض;PF*$8ap8r<0t2`6:P2-hʳʞ*(y2 *̵&מnng&mwB3R}fa]?Hd]w Yn 9UKH ~3 j>ʤ^5pP |S9$GY9R&3%sW(+˙cm wWxfڢ]_|!B;!23=&Hacgzd\ꃷr(Ю \fd~w?^}~zфzeyfxͭ![ՏǘOf.o?Bcwz ؾ쁳+rֈ'%z__̠|慷I]x)OfhWRV6Q X ])}@bY 7` ʨX"̹I5I~[ UR.QhST_6:YKOQKE-*D8 -tM5 bg-=a-{Yތi/T+YKOZK*RBRʴtMuz㟶.RYKQh)2-]SMh'eZJ3>-%(5Ղ¸T9[* 5Ղ+:k k,l{j^.yuoUvˇu9f}- t23,@cYl{{»9БXck Ѝp)o˃>PޕaW& ,[ЮBD_r־T>ÃM1LyQ?IkhJ?˦l ";8pI1m.?.{87gջ{2wEp/ZM#y˻#=pTbުE^?d|M7u>Yp26R(j0Ҏax B n| GvzJ'\T.:}V#|[:nF^ Yֶveg4U0M=B&dA+i,_ a,g9|:dL,Ff|{Pf E_.Efؐ1 2C' 48:2ä 2,{mqF\:1j@@fACu|)b!3H'#yt@pZ VgdD1 +k n+¡6 f(hAkMO#3V;{:T!TLă_M+eb}/4ZRޝ~sP ƌWi0H5xkt*WofJKGE 'epq'ZP\p9eI=zMK+j+%o؛:J%M*܍,sX Vw3JD . q)4Q yt j~wy4v)Y}zl9EIj.;N2QIÎ^xQɶJ麐^g8Dy\%QVsd w2R^EMˀmAX X"Lmヤpq<|T筐yasz`SAVP+8H!G#6B6%yl kx7R#x-*ɣUer!BPFfiS0wr6}75mڨ ]xQƂ06jczɴsx*^W'PxXLA o`pIidM]mx{uu . )@ )\Wɝ Y6KW Sp& eN5=;6ϔ^QW D\Ha3D"N8_ Pcz=KMyv#]ecJ픩eegP08gI&*pxH/C#iCz9Z:},2,s*A,r뫘>69Ss1zT _ؙf2g&FU)F# o PH܀ey,a Fy \DJE৤I ䷡l &\璱 (V5=QGŕ)] *# GbQ*s oݍXn] 1H, %g9Dg0 <9 )b2hJ*Q`UeFF8m6@[pI)Aak$/aLAogƵ Uh0 [*D1c(e#D%ӻXb4TØ%)$#%< Fubܪǚ{"=} ʇl '+gUx,"**Y.avj֩es8ͽgʟa) GgZHtjy RSZ..D!)QtVB؏ pfAڒ!\Q6"I!&Ciq48¤%c,r[oG,9]i(1hXH8G.YBT,aRnkSٷ)$'ܕ1(`)+d^R[gPLP U$ )D04p)L!eEWL:]iMe rcKSb) m8`q[ʮi/qbTxRMkҎp25p*jNL@CoD-$3{j] $LJF5IHA̭mT_S19ͭ[iס:4B8@*Ӭq|*m>n+5L3jPiNVz\V㸦T_SMS}cRӬov+~@tH ZLR%H֬F҅PL!zH#՚ŨW52$XZfsPRWZ$!+3]ۂ;sV pbTK$#B^a>& Y\1Vz\b p+J7ǥ|4ZNskomggxp֛1Nuڀ ?9x2٢ *-eS$㵌ӫڟ_|h Y[!k|ŬHW7jAH{U[^,އ jL  nxIkDm&rLt,9V{Smxk>ʔ/Ԉj[~v;}6֞O8T`>6(zNkِjr$:ZUjWdNΓzSO9(؎OO?P&Ȯus Jjw88gxc SDer| ?1N*];8?MTv}W0o/%(hx$6Ǫ~o/?:Od30R:-,96 kS^YA@]Ba(21rC Ť-DwtLJ_t1YbzLZޯxD/mPOUTGJYqTlѐ >U9V)rN:1tXEKLR)Jep!W:{\)bg`{ؠIGӗ"{嫸 f<#k<fu%m_$ ބm/%SsknOFiGɗ'Wuw F5]}ߦv: }i*m)W+Q,]Jt=5bKk;XI֖L1P57!6ZCRZKEUmH(ޒ爀`ǹ_!22$a %XNߊ+aqud@<=>eH;MgEP)_p;`uiE >^R_Ie;z ˬ$ipjRْ1TRQ`J2x$:dɌDEpD8&N+ .O w-sW|i:+ջkxXVo~?+g XŻD0$aq2zxXT5t6WaaS81l SCZjCuZZf$vrFPp *GGPM0|+]mT,y 7_,yInphQ͂fPͪ.M N~qh3g꧜w^oz^Sʺ{dC\Wj~=S/=Ê<2ϕBРignKqA5\oV:Q' Hb@aׂ5b 0 :o7n{wX\A4CWuDDȐT>,P ~-g;9<ݛ-#X-Ϟ~ʑX_/<mxq+Qɇ.}w qrX5'#4/4 ;08 IcO^(MZ#NR&,X6Up9- ִ~pVDb S+&x'dKe?%<+20_'@ }Хx_|h^8t-扱7ɃB0o8#+ӷ)eIn޷%˓,0"ЇXxV<0Q gSNd~7cnφg|1gT(l+t< eedy57'x%pmQ'c*7n"標2)~GݪЕK{+X)?!T'OHΦ\+}/5^LJU ; pAmJc ՆwxfgW5u}o]CJ28騟(,⤛~ )Ygc8ECEyq`1@nC|b7>}^Gyݵ.̭IH>0xC0HY{l)cT ޹xjcY`-EE0/_h8at>]Jٵ+Ibdd(ri-,Ϗ\!F./dv$m]b$ N.i$KBWz`:YrKH"K7+}l{h >iN u=y3 |^sz3UXMuwq~^|vr]%)T t*8 g|I3 Z$ DDu#am<%G4l+KGC2?ecԱ홹d'm`;YT# b [bT9<ԹwjفAOyܮ;X} zr޴ux1J 4G M=!=2,A,^;F,06">޹yQBNfImAeNV *WyלY4z޸EGm@hMYF?n :{n2H1wx=]"oAhM)~3& {n2H1wx}f24w+JnCXnk66ʁJkvS^) }S"Ԑ?{Wƍ/{EVr9p}KphII;cǾ)3IC6[Ҍ $qbUX$J0m4 > o2n#)l#Ud:Č6R㶑sKmMyc>1#sۀ69jFɋVlhԐ)5DŻ_ܻ!Y*`nF4&nOWʶKI4Iae.LV褽ìp|w(C6\WTYg Ç$v竚ۏ*Y8@i6G1=ѐi"2x&ϧM7PK&#:d2NZH e¦%ePh3_l?'s>Y*2PdqJ~ɇ(gi7^ t,MsFQ.]Ke*:~u2S w{I82{e.I: y0gސg-4u\jʮr|wD!06< O}5Zdig>>7ʣRmH܅Ŷ nT\M2YB§]a<Z$-QeJV;K~V-H\M8r6m2293)8NiDJ.K%?4GҩPsw ^)S!0x4ƆfVFSåS Z1I4j@S~Tqa 5T5|ȓ}\LZo0SM['c&2IWU>tt? $Xku:(n^Ѧ:vt/Ȍncx+w*RxF3FZN7JیUȱ[}AnftC^|*.M"VZLw|X= ;[ZM-D|$ۅ[ŕn!f>8,u_Z4*O a>BauA<5KYvP (CaΆt."z-6IfLuQ2@n4,>Js_/\Ye&tw{Ptw5aTCǟlvmcm9YP9|?>(KJdSo+eS_m}h>|j=\'3OoVg^7?i[?OW%Ek܇ !lm\•$ca#'8`}Y%738߶yi=_k}~ߡm @ׇ{ܖѥz"zljѨ6J˼ʊ[ږ8d-{Fj!syLԨAtkFdc@1KK%ЦMm%|ZheRlށ90T^lfύd&D4~LrkGڸGS1cH-".$"wQÍ jL2a1Dw!L.4 & #Ud,ST Pu!Xi>K)M*蕙]6f9(>M1z8= & MBG~h\ZO"ln?ߴk~u=_;?]=~PGf÷_(.Ng֓Ԃ=ڸZ%z\GwZ28UswSǔ뛛<䕻hO)q+ɓd4ctA~$͘R|V}ʌncx+w*tf4BtA~$͹@dѭfW}* S?7G_/( էc +7.b/\g}\y)@cN/=$u_j=o/祂V^Ixjom/55KwWh(݁CUE0Dؿxw3n}ho' ½Z#1jD2s"}־XYJ,xB ZÂZX=Vl6;9;x`8_l4F):a6l~izxJy_dQ/6cr.f ]m94oBhAƃVSc Zk|p8 4JivUdQhֹp}1R<%p-Dtp#)ҍ̊ ^Rq]WξL%% FNR7,%dQ̰步Kp 0 Õͺ~z> aù k1Fa]RUK&x7+p`RG,RC^Š3zRD-wͺ ??$ztUIaoW,?_l-ˌ2|<@)Sh }YA/Ak݀`{ .^m'-O,/j[mf_-y4 0c.VEWUgMITc-PrhH,/!Ca3C'"9Vb3F؈s @kwZ-Q= TIFȩˆhg&,H!P4>haC*.Uנ jH^ר>kRMQѰa!GR"ȽRB󴴤rR*dK-}&ՌQKYKU]xz.RZ:gM %)* js͂SNAjVB&GFi,ZK, XFJjKjBmeAȰflAŻ><$_DxpQjQG9" ĭ*X9PvҭnX1~Gt_zdg{7奾U~Z~WҮ?U5ȔdʅHw"mR)˅"AZ :$9Clp0 Hc鍅VlT %rSn{[@QljQ1Z(Ec58F'ipZ6kT@̀|Iu<u6Zz=0 k[Mt@|V}:Jq:jy 'Vs˛tCJxxxK,APݠ3# Bl [ u@Ƚ>vܒ 5znDY~t?rn[ѳY|ewJ|A&)t$nDEW`gIS/.ߙb$J"6,HcD\;{Վq[KO "u~zwbJ7xSNREgV2UPAh!Ǵ0-x=7k19 "H0G3{Q7Q6dCœ§lp&xHա ̩O gtN* 4hz3͐ 4A[N!|$ '!׌@@^ ʔ)G!lADb XPqZX{l*c{G4t!*A6[?x<2?z)ܝٙ«@;ߠgGr|w0%۳o@0-ٰO_6^8R /K7`,3|>XM*kpAUQ/ƣc)cozi/>fݍJN@[k*S%l)8Z5hWC3鼿ے.W Ǜ =ɰdLy(%x^J('s?C`P՛IGU+g ,_٩k.;O>/aÇ{jiB4Nx8A[L2wXxؼKwP}X 쌵]%-!a=d9riG ]M$'ăJ^MuCzztUL 94$*hόPc*sKIf1mFwvfFuXvEPDcD;(*?:1z+bAu0aKҎE%߇j(8#KY vZ 9j r#a%¼~J'z6SmCBr*#{a@=2x*-W]ϪUߪMOJNgLvƟ[ee_ԪFo/1?iFK}O~wU//e*۹򥇰J|Θ+\ r}N~EܾK,?}5QbgiW}t ڏ}u5xB֗9mۈ2}gc2h/bq6ze [_Nwn"FnQm UNECn_3ƊAd߈Zp!V.AS.z:JID7K:T}q^k H]yQo}O qy` 'R(aL.lifOJ{?.#><>$_oq s8dUӰg|*ݔ&ct:bwx#W++%3C4uYtdC vmLX룙a*j qSQ$;՞y D"ވjK(h$^E֚~M%&P$)2o+6l$|OJIh~t+jI8kh%S:}#vDs^t*_~95icZ`S8+mAuF2j;ioY缮K&p%X-2&#Tm7j$H%Đd;n68b",GRDp!oNf580f(ai-x̌:$HSA;y'h_ NqœP<C!@ +k,s'Rp\xfÐGʡAɈ\n*K\ F& *hDi_Z(m KNUİhL1(MEY b% ٻ6'U ]U}ewE^}ڊؑ8q$%GsRR<$=uuQ$'˧vcT)DZRa<.4(4qtc잫wE^|,B<sk<7dLK8)>yRܢORU6uk# 98@ U}Ni)ixzV#?jNiI?7wnluJ;D$X0G|bܓglmpA➨Wy@r 'S%{ +p%a`k@W|Pr `N΃W(H3S9L1#ʤMwAJ Z#4R3OӴz-PiN=M"o懔ӼFMM ÒdTla8u41sZoN?N~7 H%,1>Bґ('˿˨L23G .Many:vwSt󋋲/p3͘sW}$^boƱSzTE3UJ$h4]{?7p讽]35H$FHԠ@w Byw8g1 c q@+J[[!rfJ-@,p*h0y neÔQ$ϔ k5,uj'b-L Lq%|j?t کN Esi~:f[~wM6G F`g9FVW>~lܧˬDw7Xbٯen[vy P\Ml 4ĕ )"R` -x4U). g"_wJm1뷱?>_ޱ^i| #' nUI)=Q%͟2)>EV\;]{@)S5GHa@_c˙|>;k+e+2 eӭrK2;^nC8B ،}]?>Jz.}|E̔'98wNyjYEV=|oˇ|~#&8-:܃Wgn^\ɘ$õF9c7SqշԶyrR#_TYɋd%Y cl ',6d܍bU\iߩ=Z$.?-rL>! QFPCRL14hr ݟ^3wQ \[ qh/N ܡ$KA(Jy-=jeXEd,K!yԜE~`jYu6oslMIR erq 8;ƈt 󀬵W8>;B㖂)#(IߏpL+Gv=ebw;c;{)~;(r DwJҵ ZcU,;K@dn*H̚1=gEjf9P6u;"7?7hqE:׸T=\+ڐ1 FSMQYt#h_ynZ0,V [n s:]=崞$Y7\d X2~hvN_9 R9_2ia|F{ KAX*jF J4ȡPZR$-U6Iiu*6!014 ;F)C5VB!_<x bj¯?/]㺏WJQĽU ;cǠZ .?G_O;O;"Wp"xyxM3_О=΁VŠd2q|4MAF!e[O,rΘL;NweX A\ CiEE5"1qRX&B%1->-VFVٴÅ}h^" /܆*p(VP}yT"|ىhhDv\E0{48r2(-# P9A5Rk= BuhWȦ,C:ZnK^ Q]Zq@^@ t^|ɥ29c|v{s7tD^y"Zͅzi3gNr#\RA#rH(flb^0EQ5*9G_-S<}}}E?[G[\q6O/dL-Cujqaz ASݡ&ɴbX{C"PG*e#@OCXA%G1u̩ilzoMdV7$-otҬR$aw$x7;kAk_gNs>oߙ!U:$ T54aYivTi.dnYOUNBEɎw+|m׿̯v_c5EF1ВM\IFt|AiG AjGѥd:,xnVC<!&Cɼ2EҗN$! ʲAY8!6z 1X}()pLƁr\孖Bx6Q0d[]!bn43t nWn:&ӅrHi[m鯠,I+.9n4sZ{iwaan h!p(6Fnh(OR$TAoocp.vзg/.xÍ 3h9K-38vHujWJy_Uvu۳w?Ot?~^o,tq|kmz[<ϿYR*g.ןt&޳+l_v+ٗ X=~sˏvM.budm~zoIK^J%\_pH NhEm!+GH:T3WƨLҘEHR@_q~եu˳|+@A׍~/ޙ(BOj@fM%#Pσg['#Jl/ђfPcWY aNӛ}Bhx_cݜM[ÃnNE(yZ:l&4[ }\s0 OJ=}vņW\H ;V{Xc5|zẼ2yCdMk J iN?,#]F~T{vsY<>r~G+Y;JCl>[fb8ءfc"sEΖ uLqhԩ!dU0`u CASZKPlC".o8kX)D\Wts?tUu9Նsi~:ʑWfUE/Mk.VW曏thŽ..umg~:bk.יybC} pW]IF],EIGEJ"?Io;;㜆de1h,4f ^@,(Q?xRmƂ_?~)ni?̻;k2{VyQ^?N|kiVRq)i?=Z5JW^]J(j%Zs#FtY[zIs$"gPbN{#aK󻸝ֹ 5WUb_񈌺#Sp՚ql۝w"u-5j,6/ؚG=;~p ӇY 7s{m>YeAZoX"/لH2"p<1Q'j_tۃCn|^eʠLikMӘ= L6ʽmj|J 5R]%2MٳʎGzF',xKeӽlz#h@6=Ž4JaŲ9^"%aOIsH!&O ڔ 6Q\r<%V %ѣ#koË G$T.mu3 L\fTJZ]ٓ |XA;ݍK/hb%gMmz,nٷR$MeS]x!]Ϧ48CܣQʡv={Irx<7zPlp=NI 57|73 Ob~bSO ;}dnfI|ۋ|sߛO?;/cs 8jpC>iwC8ћ8֢tNO4lFe8iz'yC" yNEb¹凛P-K{ ֫v)hս_hս }_uuѪ5idh~dֿpaUnRj2=N\Ob'~z/ћu||pwbB︯><]%rٓh=j~_CL'ыcʛ=$o.d|_=#&q`ҠM F[N7^2ⱯcfZ֭/ Nĺd!gnEvnCH\Dd : ij=:?IIC w7 .T`Jp>M]WIne ܟcժ?-Z{xx9Dbd#UשUש#~̈́SMѿ3G^(֑cJ  -FQw?wjMC߉eCz7r299F$K ) %=h<0Y>_T mť٤LV$ v`H$,rPۡ ofz |8l]q2y੶\B u=g!z` ɹ̚u(w̮Oiso+'@a'%6d_{/Bҏ~wRZ!Oֈ}R,Cd'SO|{WI9*2{zUw}nT`!:1'M$Am0^@)@dd<X q$w5_fժ?=ZUWبE&֊vڗU!.ZJuGWaA.z_7s(Wnߙw1H62Ad'|EXRKggnVx:yvyՊ*MEח F$ĤJ=Q\m \`B#ayOԜOJjd̑Μy=KH^g~/J,X[u8N `k~& )WK\")#s6E(Bl`3Z<<ޅ {:H:BK5& OׂE*#^lB )L1dL ߯"E(YI@#92Yy0t<9E4^|=xфݤTv0{!vқU;!ߥUKnRpTO 4ɷ45F ALj!bZN5ڬ.a:BҪ}r_u< &YXcN(1K! AԪBf.QbT4AdWL@۽*T1JAn)-\|zSrb`Y@9L~I¡{k6v !HfʍP+a{cN{-vp\kTUzjluzyO_erRD |hq3v lSb/fP}PFA2;s{sse_(`KېS4€aמ6 #Ԓy-#Ӎ1ykn#wXk[B`vjOOJ+ \CDɑh CFS=dش}iWK\H,,zL׼tk3w)b''娡M{f Ր):.y'k4濲Rdzf] Vjy _?nԿ7SYsj'ߏ]Okսh?& ~*P}Z)WLH9|svϛ|6K8uO~Ԯ^g/a KY8jKZW%Cx'GK$akz~=/2$ggN$K YQ!j^^`5k[5bM{eߓ~s V_o4/+͟Uf&>lFM5ǟ70Dര^lIƱwCT>}:B|.yz3c>cƉFrY$~ ge#9>E?nGCoO|n"ɵd;1,Ve><0:Qr[/3ԍEO|[`rݾ].eOzW^<,wăw |Gy}~; jD׽۩0ׁlÇw?я'5atüx$z68#[:yÿe~4[lSX-Q%,@ 攲RʐY M KY,@pyf ;\4$ũ9â%䵶4\3.aBNh$U\рu( (˚!ٲRz֌%(_nT*3g Tp^jX^)bF2gրR beGL$u2r '+Ok7@1 qC/$'(fR.i*/!4~駋_.yDljX+ T"< _OjGmD5Ԧ/8'g:@j/jqd:HZ*8pB݌y9bԋ1vȤ$ j#.08\}Lx8cuᒏ0FNO;G3Dfsq6r~X?-r+RPHDl UT,d%AVd2V1*PGMOoð5(])ID*RɃ3ޱ^ %eгi+s׼^CL͞&ؔ]R7Bzɼ}$rc݃=$0heM&G;0vuLOadz5~ƤϻdNBvFײ8vy7^`|?*^?%8}=c+[ֹ8D^9DMdkgy{FQc~?_a"~љ8[1#|=Stz][ۗ>_`:["v⭻h`>VzYN'T ɑCih}VߖֺܮKXkFC Q_@]9chȄa_ N&nFC`-vf}kD吿ιĬKޭ}xb s#m"Z}{{UCC-ūP EBdu>-C-(}חTz~ƣ0gcMA${BŻ?'Ma833B(O?€~?>6('>AvAM=Y'^cd2y⦝2r>z$ ewz9"Nj$!o\D;T"r_ ֕%u&m Ti{ڭ y"E$%{tum1p4։vNKψW2V62p$y;Ysrg5ؤy2rU77B&\j5*'0NW3A5.|Tt9'Pk]D쨩rPYTM=o ]龣n6R7dD] &"[ gOO\pHZI$Z*#ytW?g5h6{saDFp5!xE*"_A.jr#% )uO㩯n؞ڇɨ tfpeH>},d26M^Z۔m#-db #`3]0= +XRNY19Kj;'}Pϴ>Xx{DB CoA~DťaURp`=~~R6MbZd{[QnE.j5?2O q'jI[ƣѤ;>'P- Fna\ގ# i%tzW`QNrm5k ^HKP-CCRV >җ>[Z7飔2V& hrOYEbu%NT ! M⬎>Њ (>)7rzkp`pM-~ex^7Vԭ EmnPggPt<9l =%NJHNIR U.<ϻ~ڛ]{]]_/d..sք<2ݝU8&uy箦NOEHC)ȹJBAUTAY-b3)jY{BW!U}&+mI3Vx06 9u}\ RjS%eN=9Քֱoׅc0{h0ޫcb+Ys7^bfOO+z7.dϴH:ҠhXS5ۡݺAB޸.T3O5LYR5LQз|+؋[闠>[Cx+}P!Wh0# )􀥔al*)o3->K\$g6y}le0]-67 U츮k ;?g$sL~_R w[הd瞲P:  7;F1J{ЂnlQWܦ@cؾ Uȭ5*l-K~txn8㲞v4nz!l0JyQiX$y`OYsUkx\ f˵iGYq9[j{Q;I8j|iTSz~Ds@V_f+%1#’lIf?n'C+{}|ÙoK`B0QUb{]TXuJ FR<&9Aj1KI,7bU#+lѴeϞAkI<Rʇ% Sdɞ]9R!($YaQu;hb>O]`]r&|p#YWݔTx5ܝɹ0yY<80Jnk>lc܏mOrt٤bQk"͞~,T)R*%G9V.VeIQF#h|o.kϜBT䐢w4{eۏw&KE`k VC 7}N$TaGC ـ%oX 0,Cr X5#f2Ԇ{cx];?7ԒדI#0 Tx X1Y$ ()wry,%Yb/{)Ԝ+~[x3#BJPs[J)R1:, );+Ӥ cs6 Z[Tp3, UDu$\u2E Z97HS8]# PY+gZw%]XTk8KQu}-VV4#5džf>]6Ϝ# h{3IER ޫK a@e}(J )UR)YfǛuT^,2 gB,fzt-5[yW"6={ #Ǡ &Wg{L Gq@iۣEiM7}4Aia?G UHZ5J??S.-CHP|rQB/fIӀ0"]ATwV!nB[tvXpNjY;+m 47;,m,@ M uߐL2Jvx8ߕ[WFHnk+R)PY95kVd6uU\0ȣvMn0K-!v]`Ncߥfٰavxr[_΋/-h:HZM= b(JCW*>S M(Ymߖ4R+xέ7j"P8"B# Z$Eb0 +0n6_aXDK~MEnXACJm9vHQ$Ħ\ #D&7lsx\ț0a:7enz[]e6E"&n|ySܞ |D*ζcМ+!we{~>]T[m|/']I5_sOsS#s e;R 2Xx; ;`ԣ.~v5ط F F/]0V"{A% FqCuނ6 F4vq SቊHКR ugVkb&o.3[aI{V6/Ճ*3E8=++3 v@'6P|nడKYxNVPg(nq/I]KB{U[AY7g zAF@#(I Fp.3Ji4li\l[[ǵw k;.0]ׇN˫|ʶ?FGk<84'"ghBLho<FGȲKϸ!z0G*/8E2k!FgN qRW(F4Q$?CD"^/rˑ;vWޱz<=CVT_2\W )'">,~FCp{|P"=?}0(xU]9$PWFBJHQ9:"ڪ.,jjpR #6#Te .'dA<]j)'/m%[dxSઑP)n6wy] `&翹~)#v[ map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 20 10:57:33 crc kubenswrapper[4846]: body: Mar 20 10:57:33 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:17.2985845 +0000 UTC m=+8.664992776,LastTimestamp:2026-03-20 10:57:17.2985845 +0000 UTC m=+8.664992776,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:33 crc kubenswrapper[4846]: > Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.479215 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e87746a39b0db openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:17.298675931 +0000 UTC m=+8.665084197,LastTimestamp:2026-03-20 10:57:17.298675931 +0000 UTC m=+8.665084197,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.487112 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 10:57:33 crc kubenswrapper[4846]: &Event{ObjectMeta:{kube-apiserver-crc.189e8775f69d7476 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 10:57:33 crc kubenswrapper[4846]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 10:57:33 crc kubenswrapper[4846]: Mar 20 10:57:33 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:23.948991606 +0000 UTC m=+15.315399852,LastTimestamp:2026-03-20 10:57:23.948991606 +0000 UTC m=+15.315399852,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:33 crc kubenswrapper[4846]: > Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.493715 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e8775f69ea8a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:23.949070498 +0000 UTC m=+15.315478734,LastTimestamp:2026-03-20 10:57:23.949070498 +0000 UTC m=+15.315478734,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.499541 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e8775f69d7476\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 10:57:33 crc kubenswrapper[4846]: &Event{ObjectMeta:{kube-apiserver-crc.189e8775f69d7476 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 10:57:33 crc kubenswrapper[4846]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 10:57:33 crc kubenswrapper[4846]: Mar 20 10:57:33 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:23.948991606 +0000 UTC m=+15.315399852,LastTimestamp:2026-03-20 10:57:23.955365034 +0000 UTC m=+15.321773270,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:33 crc kubenswrapper[4846]: > Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.506713 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e8775f69ea8a2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e8775f69ea8a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:23.949070498 +0000 UTC m=+15.315478734,LastTimestamp:2026-03-20 10:57:23.955503307 +0000 UTC m=+15.321911543,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.513501 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e87734079fd37\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e87734079fd37 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:12.303279415 +0000 UTC m=+3.669687681,LastTimestamp:2026-03-20 10:57:24.494229413 +0000 UTC m=+15.860637689,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.520072 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e87734dff995a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e87734dff995a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:12.530139482 +0000 UTC m=+3.896547718,LastTimestamp:2026-03-20 10:57:24.838344245 +0000 UTC m=+16.204752481,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.526606 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e87734eead9b9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e87734eead9b9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:12.545556921 +0000 UTC m=+3.911965157,LastTimestamp:2026-03-20 10:57:24.851175781 +0000 UTC m=+16.217584017,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.529245 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 10:57:33 crc kubenswrapper[4846]: &Event{ObjectMeta:{kube-controller-manager-crc.189e8776be501dc6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 10:57:33 crc kubenswrapper[4846]: body: Mar 20 10:57:33 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299366342 +0000 UTC m=+18.665774618,LastTimestamp:2026-03-20 10:57:27.299366342 +0000 UTC m=+18.665774618,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:33 crc kubenswrapper[4846]: > Mar 20 10:57:33 crc kubenswrapper[4846]: E0320 10:57:33.536533 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8776be51ae05 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299468805 +0000 UTC m=+18.665877081,LastTimestamp:2026-03-20 10:57:27.299468805 +0000 UTC m=+18.665877081,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:34 crc kubenswrapper[4846]: I0320 10:57:34.260206 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:35 crc kubenswrapper[4846]: I0320 10:57:35.254483 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:36 crc kubenswrapper[4846]: W0320 10:57:36.236285 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:36 crc kubenswrapper[4846]: E0320 10:57:36.236383 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:36 crc kubenswrapper[4846]: I0320 10:57:36.259241 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.259418 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.299034 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.299475 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.299681 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.300065 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.301935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.302006 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.302025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.302804 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.303160 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3" gracePeriod=30 Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.307972 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8776be501dc6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 10:57:37 crc kubenswrapper[4846]: &Event{ObjectMeta:{kube-controller-manager-crc.189e8776be501dc6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 10:57:37 crc kubenswrapper[4846]: body: Mar 20 10:57:37 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299366342 +0000 UTC m=+18.665774618,LastTimestamp:2026-03-20 10:57:37.299438801 +0000 UTC m=+28.665847047,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:37 crc kubenswrapper[4846]: > Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.316952 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8776be51ae05\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8776be51ae05 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299468805 +0000 UTC m=+18.665877081,LastTimestamp:2026-03-20 10:57:37.299618545 +0000 UTC m=+28.666026791,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.324615 4846 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e877912957a7c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:37.303132796 +0000 UTC m=+28.669541062,LastTimestamp:2026-03-20 10:57:37.303132796 +0000 UTC m=+28.669541062,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.354140 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.355785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.355840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.355858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.355933 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.362545 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.371797 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:57:37 crc kubenswrapper[4846]: W0320 10:57:37.425286 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.425512 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.434015 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8772df90443f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8772df90443f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:10.677349439 +0000 UTC m=+2.043757675,LastTimestamp:2026-03-20 10:57:37.427131514 +0000 UTC m=+28.793539790,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.553528 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.554303 4846 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3" exitCode=255 Mar 20 10:57:37 crc kubenswrapper[4846]: I0320 10:57:37.554362 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3"} Mar 20 10:57:37 crc kubenswrapper[4846]: W0320 10:57:37.590540 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.590613 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.673205 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8772f5108506\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8772f5108506 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:11.038076166 +0000 UTC m=+2.404484402,LastTimestamp:2026-03-20 10:57:37.663657196 +0000 UTC m=+29.030065462,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:37 crc kubenswrapper[4846]: E0320 10:57:37.683245 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8772f5dcd999\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8772f5dcd999 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:11.051467161 +0000 UTC m=+2.417875427,LastTimestamp:2026-03-20 10:57:37.67728035 +0000 UTC m=+29.043688586,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.259565 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.569383 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.570305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd"} Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.570572 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.572139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.572229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:38 crc kubenswrapper[4846]: I0320 10:57:38.572260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.260133 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:39 crc kubenswrapper[4846]: E0320 10:57:39.449215 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.574137 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.576216 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.576299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.576320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:39 crc kubenswrapper[4846]: I0320 10:57:39.718471 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 10:57:40 crc kubenswrapper[4846]: I0320 10:57:40.259384 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:40 crc kubenswrapper[4846]: I0320 10:57:40.577537 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:40 crc kubenswrapper[4846]: I0320 10:57:40.578941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:40 crc kubenswrapper[4846]: I0320 10:57:40.578989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:40 crc kubenswrapper[4846]: I0320 10:57:40.579008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:40 crc kubenswrapper[4846]: W0320 10:57:40.697195 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 20 10:57:40 crc kubenswrapper[4846]: E0320 10:57:40.697285 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:41 crc kubenswrapper[4846]: I0320 10:57:41.261338 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:42 crc kubenswrapper[4846]: I0320 10:57:42.260403 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:43 crc kubenswrapper[4846]: I0320 10:57:43.257713 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.257167 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.298638 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.298979 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.307549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.307624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.307643 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.321718 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.323596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.323653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.323672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.324515 4846 scope.go:117] "RemoveContainer" containerID="edfc627becf0226a7824062bdac8653b43e6a2c98003d3e2d0eebc17094e70a9" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.363764 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.366753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.366825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.366845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.366922 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:57:44 crc kubenswrapper[4846]: E0320 10:57:44.374405 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:57:44 crc kubenswrapper[4846]: E0320 10:57:44.374446 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.591603 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.593386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676"} Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.593572 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.594542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.594593 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:44 crc kubenswrapper[4846]: I0320 10:57:44.594612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.258556 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.600561 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.601574 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.606513 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" exitCode=255 Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.606654 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676"} Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.606858 4846 scope.go:117] "RemoveContainer" containerID="edfc627becf0226a7824062bdac8653b43e6a2c98003d3e2d0eebc17094e70a9" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.607563 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.611130 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.611207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.611234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:45 crc kubenswrapper[4846]: I0320 10:57:45.612423 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:57:45 crc kubenswrapper[4846]: E0320 10:57:45.612842 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:57:46 crc kubenswrapper[4846]: I0320 10:57:46.260236 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:46 crc kubenswrapper[4846]: I0320 10:57:46.612888 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 10:57:47 crc kubenswrapper[4846]: I0320 10:57:47.261797 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:47 crc kubenswrapper[4846]: I0320 10:57:47.299470 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 10:57:47 crc kubenswrapper[4846]: I0320 10:57:47.299835 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 10:57:47 crc kubenswrapper[4846]: E0320 10:57:47.309389 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8776be501dc6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 10:57:47 crc kubenswrapper[4846]: &Event{ObjectMeta:{kube-controller-manager-crc.189e8776be501dc6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 10:57:47 crc kubenswrapper[4846]: body: Mar 20 10:57:47 crc kubenswrapper[4846]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299366342 +0000 UTC m=+18.665774618,LastTimestamp:2026-03-20 10:57:47.299745699 +0000 UTC m=+38.666154005,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 10:57:47 crc kubenswrapper[4846]: > Mar 20 10:57:47 crc kubenswrapper[4846]: E0320 10:57:47.322632 4846 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e8776be51ae05\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e8776be51ae05 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 10:57:27.299468805 +0000 UTC m=+18.665877081,LastTimestamp:2026-03-20 10:57:47.299963404 +0000 UTC m=+38.666371680,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 10:57:48 crc kubenswrapper[4846]: I0320 10:57:48.260737 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.258597 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.428769 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.429058 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.430945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.431021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.431044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:49 crc kubenswrapper[4846]: I0320 10:57:49.436570 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:57:49 crc kubenswrapper[4846]: E0320 10:57:49.437282 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:57:49 crc kubenswrapper[4846]: E0320 10:57:49.450014 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:57:50 crc kubenswrapper[4846]: I0320 10:57:50.259403 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.259703 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.374535 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.376852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.376996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.377017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.377064 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:57:51 crc kubenswrapper[4846]: E0320 10:57:51.382113 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:57:51 crc kubenswrapper[4846]: E0320 10:57:51.382207 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.688825 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.689169 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.690800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.690877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.690927 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:51 crc kubenswrapper[4846]: I0320 10:57:51.692086 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:57:51 crc kubenswrapper[4846]: E0320 10:57:51.692455 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:57:52 crc kubenswrapper[4846]: I0320 10:57:52.259517 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:53 crc kubenswrapper[4846]: I0320 10:57:53.260524 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.257039 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:54 crc kubenswrapper[4846]: W0320 10:57:54.331800 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:54 crc kubenswrapper[4846]: E0320 10:57:54.331994 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.482609 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.482838 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.485005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.485070 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.485094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.488038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.643145 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.644236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.644269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:54 crc kubenswrapper[4846]: I0320 10:57:54.644279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:55 crc kubenswrapper[4846]: W0320 10:57:55.092167 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 10:57:55 crc kubenswrapper[4846]: E0320 10:57:55.092229 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 10:57:55 crc kubenswrapper[4846]: I0320 10:57:55.258050 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:56 crc kubenswrapper[4846]: I0320 10:57:56.259001 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:57 crc kubenswrapper[4846]: I0320 10:57:57.257575 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.259705 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.382582 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.384096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.384181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.384237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:57:58 crc kubenswrapper[4846]: I0320 10:57:58.384272 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:57:58 crc kubenswrapper[4846]: E0320 10:57:58.389458 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:57:58 crc kubenswrapper[4846]: E0320 10:57:58.389595 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:57:59 crc kubenswrapper[4846]: I0320 10:57:59.260506 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:57:59 crc kubenswrapper[4846]: E0320 10:57:59.450165 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:58:00 crc kubenswrapper[4846]: I0320 10:58:00.256140 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:00 crc kubenswrapper[4846]: W0320 10:58:00.468479 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 10:58:00 crc kubenswrapper[4846]: E0320 10:58:00.468576 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 10:58:01 crc kubenswrapper[4846]: I0320 10:58:01.259938 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:02 crc kubenswrapper[4846]: I0320 10:58:02.255886 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.259862 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.322267 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.324301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.324376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.324399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:03 crc kubenswrapper[4846]: I0320 10:58:03.325500 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:58:03 crc kubenswrapper[4846]: E0320 10:58:03.325863 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:03 crc kubenswrapper[4846]: W0320 10:58:03.360053 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 20 10:58:03 crc kubenswrapper[4846]: E0320 10:58:03.360145 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 10:58:04 crc kubenswrapper[4846]: I0320 10:58:04.258023 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.262840 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.390090 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.392125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.392213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.392258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.392310 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:58:05 crc kubenswrapper[4846]: E0320 10:58:05.401817 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:58:05 crc kubenswrapper[4846]: E0320 10:58:05.404704 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.801661 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.801962 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.803448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.803531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:05 crc kubenswrapper[4846]: I0320 10:58:05.803556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:06 crc kubenswrapper[4846]: I0320 10:58:06.259708 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:07 crc kubenswrapper[4846]: I0320 10:58:07.257388 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:08 crc kubenswrapper[4846]: I0320 10:58:08.259506 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:09 crc kubenswrapper[4846]: I0320 10:58:09.257932 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:09 crc kubenswrapper[4846]: E0320 10:58:09.450480 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:58:10 crc kubenswrapper[4846]: I0320 10:58:10.260087 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:11 crc kubenswrapper[4846]: I0320 10:58:11.257167 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.259559 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.402537 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.404840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.404875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.404884 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:12 crc kubenswrapper[4846]: I0320 10:58:12.404943 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:58:12 crc kubenswrapper[4846]: E0320 10:58:12.410473 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 10:58:12 crc kubenswrapper[4846]: E0320 10:58:12.411371 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 10:58:13 crc kubenswrapper[4846]: I0320 10:58:13.256287 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 10:58:13 crc kubenswrapper[4846]: I0320 10:58:13.802013 4846 csr.go:261] certificate signing request csr-6gdn4 is approved, waiting to be issued Mar 20 10:58:13 crc kubenswrapper[4846]: I0320 10:58:13.813551 4846 csr.go:257] certificate signing request csr-6gdn4 is issued Mar 20 10:58:13 crc kubenswrapper[4846]: I0320 10:58:13.870819 4846 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 20 10:58:14 crc kubenswrapper[4846]: I0320 10:58:14.099816 4846 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 20 10:58:14 crc kubenswrapper[4846]: I0320 10:58:14.815523 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-20 01:38:42.629433344 +0000 UTC Mar 20 10:58:14 crc kubenswrapper[4846]: I0320 10:58:14.815600 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6590h40m27.813839987s for next certificate rotation Mar 20 10:58:15 crc kubenswrapper[4846]: I0320 10:58:15.322015 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:15 crc kubenswrapper[4846]: I0320 10:58:15.324155 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:15 crc kubenswrapper[4846]: I0320 10:58:15.324389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:15 crc kubenswrapper[4846]: I0320 10:58:15.324545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.322100 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.323561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.323638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.323653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.324639 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.714927 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.717127 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e"} Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.717279 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.718639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.718663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:18 crc kubenswrapper[4846]: I0320 10:58:18.718672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.411883 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.414637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.414680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.414691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.414837 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.424285 4846 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.424624 4846 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.424656 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.428950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.428971 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.428980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.428993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.429005 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:19Z","lastTransitionTime":"2026-03-20T10:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.444058 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.450693 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.461189 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.461246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.461258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.461276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.461287 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:19Z","lastTransitionTime":"2026-03-20T10:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.472887 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.479724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.479746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.479756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.479775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.479787 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:19Z","lastTransitionTime":"2026-03-20T10:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.495382 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.506361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.506416 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.506427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.506447 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.506474 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:19Z","lastTransitionTime":"2026-03-20T10:58:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.521004 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.521348 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.521441 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.622351 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.722549 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.724021 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.725021 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.731257 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" exitCode=255 Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.731318 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e"} Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.731362 4846 scope.go:117] "RemoveContainer" containerID="7785ba00cc26c24d49bea0b0f2f80d98cdfcc3160cfd5fc7fc0fb04d92c97676" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.731568 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.733087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.733122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.733134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:19 crc kubenswrapper[4846]: I0320 10:58:19.733839 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.734069 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.823696 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:19 crc kubenswrapper[4846]: E0320 10:58:19.924156 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.025224 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.126416 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.227095 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.327712 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.428346 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.529265 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.630497 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.731118 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: I0320 10:58:20.736125 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.832189 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:20 crc kubenswrapper[4846]: E0320 10:58:20.932326 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.032503 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.133274 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.234384 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.335205 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.435628 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.537300 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.637658 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.689000 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.689688 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.691577 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.691658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.691678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:21 crc kubenswrapper[4846]: I0320 10:58:21.692967 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.693311 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.738801 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.839145 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:21 crc kubenswrapper[4846]: E0320 10:58:21.940202 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.041224 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.142139 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.242490 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.342800 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.443912 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.544999 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.646025 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.746448 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.847482 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:22 crc kubenswrapper[4846]: E0320 10:58:22.948077 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.049309 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.149893 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.250415 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.351013 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.452131 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.553007 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.654140 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.754268 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.854510 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:23 crc kubenswrapper[4846]: E0320 10:58:23.955573 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.055956 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.156391 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.257547 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.357810 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.458120 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.558598 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.659841 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.760012 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.860176 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:24 crc kubenswrapper[4846]: E0320 10:58:24.961280 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.061735 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.162681 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.263547 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.364179 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.465710 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.565970 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.667135 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.767635 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.868324 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:25 crc kubenswrapper[4846]: E0320 10:58:25.968779 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.069850 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.170253 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.270780 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.371648 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.472852 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: I0320 10:58:26.478008 4846 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.574057 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.674761 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.775188 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.875937 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:26 crc kubenswrapper[4846]: E0320 10:58:26.976267 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.077441 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.178040 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.279162 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.379512 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.479641 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.579961 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.680428 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.781071 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.881528 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:27 crc kubenswrapper[4846]: E0320 10:58:27.982663 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.083824 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.184452 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.285250 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.386012 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.487639 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.589746 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.691147 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.791706 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.892866 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:28 crc kubenswrapper[4846]: E0320 10:58:28.993116 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.093314 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.193707 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.294550 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.351193 4846 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.395634 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.428096 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.428599 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.429798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.429831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.429841 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.430419 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.430575 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.451953 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.496289 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.597425 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.690706 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.717637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.717704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.717744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.717781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.717808 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:29Z","lastTransitionTime":"2026-03-20T10:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.731716 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.736753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.736785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.736796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.736814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.736828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:29Z","lastTransitionTime":"2026-03-20T10:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.752711 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.760058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.760130 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.760148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.760175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.760195 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:29Z","lastTransitionTime":"2026-03-20T10:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.775564 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.783100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.783172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.783194 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.783587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:29 crc kubenswrapper[4846]: I0320 10:58:29.783798 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:29Z","lastTransitionTime":"2026-03-20T10:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.800715 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.800960 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.800996 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:29 crc kubenswrapper[4846]: E0320 10:58:29.901583 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.002773 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.103749 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.204462 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.305470 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.406360 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.507107 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.607529 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.708143 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.808351 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:30 crc kubenswrapper[4846]: E0320 10:58:30.908855 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.009085 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.109687 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.210325 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.310656 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.410870 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.511475 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.612525 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.712818 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.813447 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:31 crc kubenswrapper[4846]: E0320 10:58:31.914296 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.015020 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.115609 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.216262 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.316452 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.416938 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.517174 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.617688 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.718294 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.818475 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:32 crc kubenswrapper[4846]: E0320 10:58:32.919247 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.020190 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.121265 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.222086 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.322785 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.423249 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.523673 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.624163 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.724583 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.825119 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:33 crc kubenswrapper[4846]: E0320 10:58:33.925355 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.025560 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.125736 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.226383 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.327525 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.428667 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.528850 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.629096 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.729924 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.830689 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:34 crc kubenswrapper[4846]: E0320 10:58:34.931463 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.032529 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.132678 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.233276 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.333720 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.434559 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.535678 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.636174 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.737360 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.838259 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:35 crc kubenswrapper[4846]: E0320 10:58:35.938796 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.039128 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.139542 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.240108 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.341243 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.442167 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.542785 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.643117 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.743304 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.843876 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:36 crc kubenswrapper[4846]: E0320 10:58:36.944313 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.045267 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.146388 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.246718 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: I0320 10:58:37.321738 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 10:58:37 crc kubenswrapper[4846]: I0320 10:58:37.323286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:37 crc kubenswrapper[4846]: I0320 10:58:37.323378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:37 crc kubenswrapper[4846]: I0320 10:58:37.323399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.347457 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.448385 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.548744 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.649538 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.749888 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.851062 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:37 crc kubenswrapper[4846]: E0320 10:58:37.951407 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.052422 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.153499 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.254359 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.354526 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.455204 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.555633 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.656785 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.757479 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.858440 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:38 crc kubenswrapper[4846]: E0320 10:58:38.959475 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.060421 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.161444 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.185605 4846 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.264974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.265047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.265065 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.265093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.265113 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.280816 4846 apiserver.go:52] "Watching apiserver" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.289432 4846 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.290363 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqml5","openshift-multus/multus-additional-cni-plugins-z8m7m","openshift-multus/network-metrics-daemon-zljxc","openshift-image-registry/node-ca-7rtn2","openshift-machine-config-operator/machine-config-daemon-jds6p","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl","openshift-dns/node-resolver-j4h4w","openshift-multus/multus-cqnfg","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.291021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.291247 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.291368 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.291585 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.291671 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.291993 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.292203 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.292124 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.292354 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.292674 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.293539 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.293627 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.293691 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.294403 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.294586 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.294611 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.294918 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.295137 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.296698 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297170 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297332 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297359 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297477 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297500 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297560 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297681 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297757 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297767 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.297885 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.299029 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.299164 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.301881 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.307156 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.307216 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.307424 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.307817 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.307871 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.308095 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.308369 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.308795 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309172 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309401 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309565 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309618 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309868 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.310248 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.310595 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.311261 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.312294 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.312463 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309409 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.309862 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.313470 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.315059 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.316970 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.330075 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.345711 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.355253 4846 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.359940 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.368618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.368665 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.368680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.368698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.368711 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.372916 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.388459 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.400658 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.419393 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.435208 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.436728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.436811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.436852 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.436889 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.436971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437016 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437053 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437109 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437160 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437194 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437234 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437279 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437313 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437382 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437420 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437468 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437507 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437656 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437724 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.437964 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438001 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438032 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438115 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438185 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438270 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438375 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438415 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438467 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438500 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438582 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438619 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438659 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438730 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438804 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438947 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.438973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439018 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439101 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439197 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439236 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439273 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439348 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439383 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439463 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439485 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439634 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439679 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439581 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439658 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439657 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.439695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.440142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.440243 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443472 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441596 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441526 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441697 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441729 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.441778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442248 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442523 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442632 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442702 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442792 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.442790 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443095 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443708 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443856 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.443975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444008 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444075 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444089 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444111 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444429 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444438 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444488 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444431 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444557 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444592 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444621 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444758 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444842 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444870 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444912 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444924 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.444970 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445081 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445421 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445249 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445533 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445551 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445561 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445642 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.445660 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446144 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446220 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446103 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446154 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446772 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446966 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.447285 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.447482 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.447496 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448613 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.447634 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.447651 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448069 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448494 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448599 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.446295 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448890 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.448979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449058 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449107 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449165 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449203 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449239 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449312 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449361 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449503 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449584 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449622 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449660 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449706 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449760 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449868 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449956 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.449997 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450032 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450069 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450105 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450141 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450247 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450285 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450363 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450406 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450502 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450553 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450598 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450706 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450772 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450838 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450951 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.450989 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451024 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451059 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451129 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451234 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451299 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451368 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451446 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451545 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451736 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.451791 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452009 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452064 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452101 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452359 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452253 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452760 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452808 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452830 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452876 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453012 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453057 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453103 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453123 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453186 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453221 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453298 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453315 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453355 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453377 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453454 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453550 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453604 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453622 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453682 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453752 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzlmb\" (UniqueName: \"kubernetes.io/projected/b2799b9e-c7c5-4816-97d6-d181a459f81f-kube-api-access-xzlmb\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453880 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453966 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454004 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454024 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-os-release\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454087 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-system-cni-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-bin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454195 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454214 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454247 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454287 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv8k6\" (UniqueName: \"kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454406 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-system-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b2799b9e-c7c5-4816-97d6-d181a459f81f-host\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-hostroot\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454819 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzg6v\" (UniqueName: \"kubernetes.io/projected/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-kube-api-access-bzg6v\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-k8s-cni-cncf-io\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455030 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-multus\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msz4s\" (UniqueName: \"kubernetes.io/projected/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-kube-api-access-msz4s\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455216 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b2799b9e-c7c5-4816-97d6-d181a459f81f-serviceca\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cnibin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-etc-kubernetes\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455403 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-rootfs\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455446 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455472 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455543 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-kubelet\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-multus-certs\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455632 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-proxy-tls\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-hosts-file\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455702 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455731 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-socket-dir-parent\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-conf-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456033 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-mcd-auth-proxy-config\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456102 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-netns\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456181 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-os-release\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456201 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456221 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-binary-copy\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456264 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456423 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cni-binary-copy\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456449 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456478 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456517 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456540 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-daemon-config\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456598 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnzt6\" (UniqueName: \"kubernetes.io/projected/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-kube-api-access-cnzt6\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456623 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2w8\" (UniqueName: \"kubernetes.io/projected/aee42334-a152-4ac7-8778-a1bdfad832af-kube-api-access-tx2w8\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzkbs\" (UniqueName: \"kubernetes.io/projected/d8533b04-cc98-4712-95b3-1dee5598b9b3-kube-api-access-xzkbs\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.458188 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.459515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.459559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cplvm\" (UniqueName: \"kubernetes.io/projected/dfe48192-f5b6-4ba5-a4a4-475459fd1815-kube-api-access-cplvm\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460757 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-cnibin\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461155 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461561 4846 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461620 4846 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461647 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461674 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461700 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461725 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461749 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452804 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461774 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452837 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.452968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461792 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453176 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453706 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453961 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453337 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.453937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454066 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454189 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454273 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454311 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.454816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455078 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455417 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462234 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461804 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462391 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462427 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462455 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462617 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.455466 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.462870 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.462935 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456858 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457097 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457250 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457362 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457764 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.457867 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.458488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.458625 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.459001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.459377 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460186 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460675 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.460745 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463649 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.461144 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.463024 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:39.962966134 +0000 UTC m=+91.329374370 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463943 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.463977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.456377 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.464085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.464155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.465065 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.465176 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:39.965151625 +0000 UTC m=+91.331559861 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.465762 4846 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.466920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.467222 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.468142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.468207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.468567 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.468653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.469041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.469267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.470156 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.471452 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:58:39.97140927 +0000 UTC m=+91.337817546 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471668 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471727 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471756 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471780 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471848 4846 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471869 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471892 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471957 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.471979 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472000 4846 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472023 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472090 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472112 4846 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472135 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472156 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472176 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472199 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472601 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472639 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472663 4846 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472686 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472706 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472726 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472748 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472768 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472802 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472822 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472842 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472863 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472882 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472936 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472956 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.472984 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473006 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473025 4846 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473045 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473065 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473087 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473107 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473127 4846 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473147 4846 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.473167 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.474197 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.474262 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.474934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.475337 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.475395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.476641 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.476724 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477375 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477459 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.477640 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.486877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.488817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.488802 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.488966 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.489339 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.489698 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.489863 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490215 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490358 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490417 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490442 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490472 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490520 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490756 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.490987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.491471 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.491532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.491501 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.491664 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.491851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492037 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492213 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492391 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492786 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.492858 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.493118 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.494488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.495630 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.495747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496102 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496222 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496496 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496603 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496642 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.496868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.497120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.497174 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.497331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.497425 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.497452 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.497470 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.497557 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:39.997532647 +0000 UTC m=+91.363940883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.497555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.498389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.498779 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.498928 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.499360 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.499394 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.499417 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.499496 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:39.999466082 +0000 UTC m=+91.365874318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.500408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.502505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.502620 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.502995 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.503709 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.509151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.511535 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.512391 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.516355 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.516908 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.517117 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.517624 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.518052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.518557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.519695 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.520298 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.520463 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.521172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.521499 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.522499 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.522582 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.522886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.522997 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.523149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.523172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.523476 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.523549 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.523702 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.526290 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.527256 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.527284 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.527491 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.527330 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.531768 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.535881 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.543110 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.544542 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.548476 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.555143 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.556186 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.565370 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574320 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cplvm\" (UniqueName: \"kubernetes.io/projected/dfe48192-f5b6-4ba5-a4a4-475459fd1815-kube-api-access-cplvm\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574363 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574391 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-cnibin\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzkbs\" (UniqueName: \"kubernetes.io/projected/d8533b04-cc98-4712-95b3-1dee5598b9b3-kube-api-access-xzkbs\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzlmb\" (UniqueName: \"kubernetes.io/projected/b2799b9e-c7c5-4816-97d6-d181a459f81f-kube-api-access-xzlmb\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574582 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-cnibin\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574802 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.574607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575221 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-os-release\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-system-cni-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575278 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575414 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-bin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-system-cni-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.575939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-os-release\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-bin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576263 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv8k6\" (UniqueName: \"kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-system-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-hostroot\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576521 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b2799b9e-c7c5-4816-97d6-d181a459f81f-host\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576583 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzg6v\" (UniqueName: \"kubernetes.io/projected/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-kube-api-access-bzg6v\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576612 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576621 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576708 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-multus\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576378 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-system-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576674 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-cni-multus\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576768 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msz4s\" (UniqueName: \"kubernetes.io/projected/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-kube-api-access-msz4s\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576810 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-hostroot\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576848 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.576854 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b2799b9e-c7c5-4816-97d6-d181a459f81f-host\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577126 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577266 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-k8s-cni-cncf-io\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-etc-kubernetes\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577301 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-rootfs\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577424 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b2799b9e-c7c5-4816-97d6-d181a459f81f-serviceca\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cnibin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577553 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-multus-certs\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577574 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-proxy-tls\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577595 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-hosts-file\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-kubelet\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577676 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-conf-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577710 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-mcd-auth-proxy-config\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cnibin\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.577853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-multus-certs\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.578387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.578526 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-k8s-cni-cncf-io\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.578766 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-conf-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.578853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-var-lib-kubelet\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-rootfs\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579148 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-etc-kubernetes\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579308 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-socket-dir-parent\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-netns\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-os-release\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579682 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cni-binary-copy\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579693 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-host-run-netns\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579733 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-binary-copy\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579784 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-daemon-config\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579809 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnzt6\" (UniqueName: \"kubernetes.io/projected/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-kube-api-access-cnzt6\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579835 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579858 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-cni-dir\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2w8\" (UniqueName: \"kubernetes.io/projected/aee42334-a152-4ac7-8778-a1bdfad832af-kube-api-access-tx2w8\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579945 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579987 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8533b04-cc98-4712-95b3-1dee5598b9b3-os-release\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.579995 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-hosts-file\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.580164 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.580228 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:58:40.080204987 +0000 UTC m=+91.446613233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-socket-dir-parent\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580448 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580450 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-multus-daemon-config\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8533b04-cc98-4712-95b3-1dee5598b9b3-cni-binary-copy\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580706 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580755 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580870 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dfe48192-f5b6-4ba5-a4a4-475459fd1815-cni-binary-copy\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.580967 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581513 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581529 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581542 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581553 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581564 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581574 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581585 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581597 4846 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581608 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581618 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581629 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581640 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581649 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581660 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581670 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581680 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581689 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581698 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581708 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581719 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581729 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581739 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581749 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581760 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581772 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581783 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581792 4846 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581802 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581815 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581827 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581839 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581851 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581862 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581873 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581882 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581910 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581925 4846 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581937 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581948 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581958 4846 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581976 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.582327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b2799b9e-c7c5-4816-97d6-d181a459f81f-serviceca\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.581985 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584174 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584237 4846 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584305 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584407 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584473 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584527 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584584 4846 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584638 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584695 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584748 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584801 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584853 4846 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.584938 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585000 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585052 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585104 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585184 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585237 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585296 4846 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585372 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585444 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587107 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.585668 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587153 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587176 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587191 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587207 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587222 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587238 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587023 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587253 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587284 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587308 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587323 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587340 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587359 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587375 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587393 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587409 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587425 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587442 4846 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587460 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587477 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587501 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587517 4846 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587533 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587549 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587567 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587587 4846 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587602 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587618 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587633 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587649 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587663 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587680 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587694 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587710 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587726 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587741 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587756 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587770 4846 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587785 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587802 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587817 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587830 4846 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587845 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587861 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587876 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587891 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587924 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587940 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587957 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587972 4846 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.587987 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588002 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588019 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588033 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588063 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588077 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588092 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588108 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588123 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588136 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588152 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588196 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588211 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588225 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588239 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588252 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588265 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588279 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588295 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588310 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.588326 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.589572 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-mcd-auth-proxy-config\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.591472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzkbs\" (UniqueName: \"kubernetes.io/projected/d8533b04-cc98-4712-95b3-1dee5598b9b3-kube-api-access-xzkbs\") pod \"multus-additional-cni-plugins-z8m7m\" (UID: \"d8533b04-cc98-4712-95b3-1dee5598b9b3\") " pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.591759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.593699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.593722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.593733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.593754 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.593771 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.595087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-proxy-tls\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.596588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msz4s\" (UniqueName: \"kubernetes.io/projected/9275b8c7-ce16-4c4a-a7eb-93d263b751d6-kube-api-access-msz4s\") pod \"node-resolver-j4h4w\" (UID: \"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\") " pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.597667 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzg6v\" (UniqueName: \"kubernetes.io/projected/aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b-kube-api-access-bzg6v\") pod \"machine-config-daemon-jds6p\" (UID: \"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\") " pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.598388 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.602189 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnzt6\" (UniqueName: \"kubernetes.io/projected/ef1b0cac-65d1-4f15-84f8-a906fa5c2967-kube-api-access-cnzt6\") pod \"ovnkube-control-plane-749d76644c-tpdzl\" (UID: \"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.602983 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv8k6\" (UniqueName: \"kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6\") pod \"ovnkube-node-cqml5\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.603062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2w8\" (UniqueName: \"kubernetes.io/projected/aee42334-a152-4ac7-8778-a1bdfad832af-kube-api-access-tx2w8\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.603516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cplvm\" (UniqueName: \"kubernetes.io/projected/dfe48192-f5b6-4ba5-a4a4-475459fd1815-kube-api-access-cplvm\") pod \"multus-cqnfg\" (UID: \"dfe48192-f5b6-4ba5-a4a4-475459fd1815\") " pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.604938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzlmb\" (UniqueName: \"kubernetes.io/projected/b2799b9e-c7c5-4816-97d6-d181a459f81f-kube-api-access-xzlmb\") pod \"node-ca-7rtn2\" (UID: \"b2799b9e-c7c5-4816-97d6-d181a459f81f\") " pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.610937 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.612256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.619794 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.628404 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.631142 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.631715 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: source /etc/kubernetes/apiserver-url.env Mar 20 10:58:39 crc kubenswrapper[4846]: else Mar 20 10:58:39 crc kubenswrapper[4846]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 10:58:39 crc kubenswrapper[4846]: exit 1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.633753 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.636175 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.643645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.644826 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 10:58:39 crc kubenswrapper[4846]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 10:58:39 crc kubenswrapper[4846]: ho_enable="--enable-hybrid-overlay" Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 10:58:39 crc kubenswrapper[4846]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 10:58:39 crc kubenswrapper[4846]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-host=127.0.0.1 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-port=9743 \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ho_enable} \ Mar 20 10:58:39 crc kubenswrapper[4846]: --enable-interconnect \ Mar 20 10:58:39 crc kubenswrapper[4846]: --disable-approver \ Mar 20 10:58:39 crc kubenswrapper[4846]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --wait-for-kubernetes-api=200s \ Mar 20 10:58:39 crc kubenswrapper[4846]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel="${LOGLEVEL}" Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.649383 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --disable-webhook \ Mar 20 10:58:39 crc kubenswrapper[4846]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel="${LOGLEVEL}" Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.651220 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.652742 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.654787 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7rtn2" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.661949 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.663268 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 10:58:39 crc kubenswrapper[4846]: W0320 10:58:39.664650 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2799b9e_c7c5_4816_97d6_d181a459f81f.slice/crio-84c0724174609fbefa1952bcd143f90124120300f93a029379880d2e06fb98fc WatchSource:0}: Error finding container 84c0724174609fbefa1952bcd143f90124120300f93a029379880d2e06fb98fc: Status 404 returned error can't find the container with id 84c0724174609fbefa1952bcd143f90124120300f93a029379880d2e06fb98fc Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.665155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j4h4w" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.665553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.666331 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 20 10:58:39 crc kubenswrapper[4846]: while [ true ]; Mar 20 10:58:39 crc kubenswrapper[4846]: do Mar 20 10:58:39 crc kubenswrapper[4846]: for f in $(ls /tmp/serviceca); do Mar 20 10:58:39 crc kubenswrapper[4846]: echo $f Mar 20 10:58:39 crc kubenswrapper[4846]: ca_file_path="/tmp/serviceca/${f}" Mar 20 10:58:39 crc kubenswrapper[4846]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 20 10:58:39 crc kubenswrapper[4846]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 20 10:58:39 crc kubenswrapper[4846]: if [ -e "${reg_dir_path}" ]; then Mar 20 10:58:39 crc kubenswrapper[4846]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: else Mar 20 10:58:39 crc kubenswrapper[4846]: mkdir $reg_dir_path Mar 20 10:58:39 crc kubenswrapper[4846]: cp $ca_file_path $reg_dir_path/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: for d in $(ls /etc/docker/certs.d); do Mar 20 10:58:39 crc kubenswrapper[4846]: echo $d Mar 20 10:58:39 crc kubenswrapper[4846]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 20 10:58:39 crc kubenswrapper[4846]: reg_conf_path="/tmp/serviceca/${dp}" Mar 20 10:58:39 crc kubenswrapper[4846]: if [ ! -e "${reg_conf_path}" ]; then Mar 20 10:58:39 crc kubenswrapper[4846]: rm -rf /etc/docker/certs.d/$d Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait ${!} Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzlmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-7rtn2_openshift-image-registry(b2799b9e-c7c5-4816-97d6-d181a459f81f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.669042 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-7rtn2" podUID="b2799b9e-c7c5-4816-97d6-d181a459f81f" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.676735 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.676751 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.689372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.689785 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -uo pipefail Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 20 10:58:39 crc kubenswrapper[4846]: HOSTS_FILE="/etc/hosts" Mar 20 10:58:39 crc kubenswrapper[4846]: TEMP_FILE="/etc/hosts.tmp" Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Make a temporary file with the old hosts file's attributes. Mar 20 10:58:39 crc kubenswrapper[4846]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo "Failed to preserve hosts file. Exiting." Mar 20 10:58:39 crc kubenswrapper[4846]: exit 1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: while true; do Mar 20 10:58:39 crc kubenswrapper[4846]: declare -A svc_ips Mar 20 10:58:39 crc kubenswrapper[4846]: for svc in "${services[@]}"; do Mar 20 10:58:39 crc kubenswrapper[4846]: # Fetch service IP from cluster dns if present. We make several tries Mar 20 10:58:39 crc kubenswrapper[4846]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 20 10:58:39 crc kubenswrapper[4846]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 20 10:58:39 crc kubenswrapper[4846]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 20 10:58:39 crc kubenswrapper[4846]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 20 10:58:39 crc kubenswrapper[4846]: for i in ${!cmds[*]} Mar 20 10:58:39 crc kubenswrapper[4846]: do Mar 20 10:58:39 crc kubenswrapper[4846]: ips=($(eval "${cmds[i]}")) Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: svc_ips["${svc}"]="${ips[@]}" Mar 20 10:58:39 crc kubenswrapper[4846]: break Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Update /etc/hosts only if we get valid service IPs Mar 20 10:58:39 crc kubenswrapper[4846]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 20 10:58:39 crc kubenswrapper[4846]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 20 10:58:39 crc kubenswrapper[4846]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 20 10:58:39 crc kubenswrapper[4846]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: continue Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Append resolver entries for services Mar 20 10:58:39 crc kubenswrapper[4846]: rc=0 Mar 20 10:58:39 crc kubenswrapper[4846]: for svc in "${!svc_ips[@]}"; do Mar 20 10:58:39 crc kubenswrapper[4846]: for ip in ${svc_ips[${svc}]}; do Mar 20 10:58:39 crc kubenswrapper[4846]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ $rc -ne 0 ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: continue Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 20 10:58:39 crc kubenswrapper[4846]: # Replace /etc/hosts with our modified version if needed Mar 20 10:58:39 crc kubenswrapper[4846]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 20 10:58:39 crc kubenswrapper[4846]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: unset svc_ips Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msz4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-j4h4w_openshift-dns(9275b8c7-ce16-4c4a-a7eb-93d263b751d6): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.691061 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-j4h4w" podUID="9275b8c7-ce16-4c4a-a7eb-93d263b751d6" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.692049 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697178 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.697745 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.706000 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-cqnfg" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.706670 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzg6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.709867 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.712206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.713741 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzg6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.714972 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.718337 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -euo pipefail Mar 20 10:58:39 crc kubenswrapper[4846]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 20 10:58:39 crc kubenswrapper[4846]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 20 10:58:39 crc kubenswrapper[4846]: # As the secret mount is optional we must wait for the files to be present. Mar 20 10:58:39 crc kubenswrapper[4846]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 20 10:58:39 crc kubenswrapper[4846]: TS=$(date +%s) Mar 20 10:58:39 crc kubenswrapper[4846]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 20 10:58:39 crc kubenswrapper[4846]: HAS_LOGGED_INFO=0 Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: log_missing_certs(){ Mar 20 10:58:39 crc kubenswrapper[4846]: CUR_TS=$(date +%s) Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 20 10:58:39 crc kubenswrapper[4846]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 20 10:58:39 crc kubenswrapper[4846]: HAS_LOGGED_INFO=1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: } Mar 20 10:58:39 crc kubenswrapper[4846]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 20 10:58:39 crc kubenswrapper[4846]: log_missing_certs Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 5 Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/kube-rbac-proxy \ Mar 20 10:58:39 crc kubenswrapper[4846]: --logtostderr \ Mar 20 10:58:39 crc kubenswrapper[4846]: --secure-listen-address=:9108 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --upstream=http://127.0.0.1:29108/ \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-private-key-file=${TLS_PK} \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-cert-file=${TLS_CERT} Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnzt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-tpdzl_openshift-ovn-kubernetes(ef1b0cac-65d1-4f15-84f8-a906fa5c2967): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: W0320 10:58:39.719280 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1a81831_258d_4fad_adc4_02dee4e744d7.slice/crio-1f8b167f2df64ee51562c3a542bd0eb419c5ab206c41a98643f31e50925f9ed3 WatchSource:0}: Error finding container 1f8b167f2df64ee51562c3a542bd0eb419c5ab206c41a98643f31e50925f9ed3: Status 404 returned error can't find the container with id 1f8b167f2df64ee51562c3a542bd0eb419c5ab206c41a98643f31e50925f9ed3 Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.722834 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.725684 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 20 10:58:39 crc kubenswrapper[4846]: apiVersion: v1 Mar 20 10:58:39 crc kubenswrapper[4846]: clusters: Mar 20 10:58:39 crc kubenswrapper[4846]: - cluster: Mar 20 10:58:39 crc kubenswrapper[4846]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: server: https://api-int.crc.testing:6443 Mar 20 10:58:39 crc kubenswrapper[4846]: name: default-cluster Mar 20 10:58:39 crc kubenswrapper[4846]: contexts: Mar 20 10:58:39 crc kubenswrapper[4846]: - context: Mar 20 10:58:39 crc kubenswrapper[4846]: cluster: default-cluster Mar 20 10:58:39 crc kubenswrapper[4846]: namespace: default Mar 20 10:58:39 crc kubenswrapper[4846]: user: default-auth Mar 20 10:58:39 crc kubenswrapper[4846]: name: default-context Mar 20 10:58:39 crc kubenswrapper[4846]: current-context: default-context Mar 20 10:58:39 crc kubenswrapper[4846]: kind: Config Mar 20 10:58:39 crc kubenswrapper[4846]: preferences: {} Mar 20 10:58:39 crc kubenswrapper[4846]: users: Mar 20 10:58:39 crc kubenswrapper[4846]: - name: default-auth Mar 20 10:58:39 crc kubenswrapper[4846]: user: Mar 20 10:58:39 crc kubenswrapper[4846]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 10:58:39 crc kubenswrapper[4846]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 10:58:39 crc kubenswrapper[4846]: EOF Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bv8k6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.725832 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_join_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_join_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_transit_switch_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_transit_switch_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: dns_name_resolver_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "false" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: persistent_ips_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "true" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # This is needed so that converting clusters from GA to TP Mar 20 10:58:39 crc kubenswrapper[4846]: # will rollout control plane pods as well Mar 20 10:58:39 crc kubenswrapper[4846]: network_segmentation_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: multi_network_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "true" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: multi_network_enabled_flag="--enable-multi-network" Mar 20 10:58:39 crc kubenswrapper[4846]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube \ Mar 20 10:58:39 crc kubenswrapper[4846]: --enable-interconnect \ Mar 20 10:58:39 crc kubenswrapper[4846]: --init-cluster-manager "${K8S_NODE}" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-bind-address "127.0.0.1:29108" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-enable-pprof \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-enable-config-duration \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v4_join_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v6_join_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${dns_name_resolver_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${persistent_ips_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${multi_network_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${network_segmentation_enabled_flag} Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnzt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-tpdzl_openshift-ovn-kubernetes(ef1b0cac-65d1-4f15-84f8-a906fa5c2967): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.729144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" podUID="ef1b0cac-65d1-4f15-84f8-a906fa5c2967" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.729243 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.734015 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: W0320 10:58:39.738526 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe48192_f5b6_4ba5_a4a4_475459fd1815.slice/crio-2a41a2584a16d6113f21eb05326eb10e12fb98db8950aa8e977f81f84194931a WatchSource:0}: Error finding container 2a41a2584a16d6113f21eb05326eb10e12fb98db8950aa8e977f81f84194931a: Status 404 returned error can't find the container with id 2a41a2584a16d6113f21eb05326eb10e12fb98db8950aa8e977f81f84194931a Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.744322 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 20 10:58:39 crc kubenswrapper[4846]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 20 10:58:39 crc kubenswrapper[4846]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cplvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-cqnfg_openshift-multus(dfe48192-f5b6-4ba5-a4a4-475459fd1815): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.744391 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.745449 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-cqnfg" podUID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" Mar 20 10:58:39 crc kubenswrapper[4846]: W0320 10:58:39.745821 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8533b04_cc98_4712_95b3_1dee5598b9b3.slice/crio-b233e326f34eb336da5d2c30859a63c18254432e34f98ad9e1ef6597734b2edc WatchSource:0}: Error finding container b233e326f34eb336da5d2c30859a63c18254432e34f98ad9e1ef6597734b2edc: Status 404 returned error can't find the container with id b233e326f34eb336da5d2c30859a63c18254432e34f98ad9e1ef6597734b2edc Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.758412 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzkbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-z8m7m_openshift-multus(d8533b04-cc98-4712-95b3-1dee5598b9b3): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.759653 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" podUID="d8533b04-cc98-4712-95b3-1dee5598b9b3" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799157 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerStarted","Data":"b233e326f34eb336da5d2c30859a63c18254432e34f98ad9e1ef6597734b2edc"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.799536 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.800379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"aefd03b98715f72e4b27a732a046a05368b833d1a11ef3212f8f8ef7422aea95"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.801538 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7rtn2" event={"ID":"b2799b9e-c7c5-4816-97d6-d181a459f81f","Type":"ContainerStarted","Data":"84c0724174609fbefa1952bcd143f90124120300f93a029379880d2e06fb98fc"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.801574 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzkbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-z8m7m_openshift-multus(d8533b04-cc98-4712-95b3-1dee5598b9b3): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.801959 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: source /etc/kubernetes/apiserver-url.env Mar 20 10:58:39 crc kubenswrapper[4846]: else Mar 20 10:58:39 crc kubenswrapper[4846]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 10:58:39 crc kubenswrapper[4846]: exit 1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.802683 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" podUID="d8533b04-cc98-4712-95b3-1dee5598b9b3" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.803021 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.803280 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 20 10:58:39 crc kubenswrapper[4846]: while [ true ]; Mar 20 10:58:39 crc kubenswrapper[4846]: do Mar 20 10:58:39 crc kubenswrapper[4846]: for f in $(ls /tmp/serviceca); do Mar 20 10:58:39 crc kubenswrapper[4846]: echo $f Mar 20 10:58:39 crc kubenswrapper[4846]: ca_file_path="/tmp/serviceca/${f}" Mar 20 10:58:39 crc kubenswrapper[4846]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 20 10:58:39 crc kubenswrapper[4846]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 20 10:58:39 crc kubenswrapper[4846]: if [ -e "${reg_dir_path}" ]; then Mar 20 10:58:39 crc kubenswrapper[4846]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: else Mar 20 10:58:39 crc kubenswrapper[4846]: mkdir $reg_dir_path Mar 20 10:58:39 crc kubenswrapper[4846]: cp $ca_file_path $reg_dir_path/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: for d in $(ls /etc/docker/certs.d); do Mar 20 10:58:39 crc kubenswrapper[4846]: echo $d Mar 20 10:58:39 crc kubenswrapper[4846]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 20 10:58:39 crc kubenswrapper[4846]: reg_conf_path="/tmp/serviceca/${dp}" Mar 20 10:58:39 crc kubenswrapper[4846]: if [ ! -e "${reg_conf_path}" ]; then Mar 20 10:58:39 crc kubenswrapper[4846]: rm -rf /etc/docker/certs.d/$d Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait ${!} Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzlmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-7rtn2_openshift-image-registry(b2799b9e-c7c5-4816-97d6-d181a459f81f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.803371 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"350ef03dc90d6738f69324bc3c61123bf64010168546668922cff852645343a7"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.804750 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-7rtn2" podUID="b2799b9e-c7c5-4816-97d6-d181a459f81f" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.805104 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.806391 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.809682 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j4h4w" event={"ID":"9275b8c7-ce16-4c4a-a7eb-93d263b751d6","Type":"ContainerStarted","Data":"dd7c870d1ca7418dbfa6f95937e1fe5b898779f1fd4c1ed9c30c34daa46ffe17"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.811808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"3939454f44b487bf87caed84ebd2ebdad9f148f723fb4181b3bdc6374f6a98ad"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.812659 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -uo pipefail Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 20 10:58:39 crc kubenswrapper[4846]: HOSTS_FILE="/etc/hosts" Mar 20 10:58:39 crc kubenswrapper[4846]: TEMP_FILE="/etc/hosts.tmp" Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Make a temporary file with the old hosts file's attributes. Mar 20 10:58:39 crc kubenswrapper[4846]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo "Failed to preserve hosts file. Exiting." Mar 20 10:58:39 crc kubenswrapper[4846]: exit 1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: while true; do Mar 20 10:58:39 crc kubenswrapper[4846]: declare -A svc_ips Mar 20 10:58:39 crc kubenswrapper[4846]: for svc in "${services[@]}"; do Mar 20 10:58:39 crc kubenswrapper[4846]: # Fetch service IP from cluster dns if present. We make several tries Mar 20 10:58:39 crc kubenswrapper[4846]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 20 10:58:39 crc kubenswrapper[4846]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 20 10:58:39 crc kubenswrapper[4846]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 20 10:58:39 crc kubenswrapper[4846]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 20 10:58:39 crc kubenswrapper[4846]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 20 10:58:39 crc kubenswrapper[4846]: for i in ${!cmds[*]} Mar 20 10:58:39 crc kubenswrapper[4846]: do Mar 20 10:58:39 crc kubenswrapper[4846]: ips=($(eval "${cmds[i]}")) Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: svc_ips["${svc}"]="${ips[@]}" Mar 20 10:58:39 crc kubenswrapper[4846]: break Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Update /etc/hosts only if we get valid service IPs Mar 20 10:58:39 crc kubenswrapper[4846]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 20 10:58:39 crc kubenswrapper[4846]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 20 10:58:39 crc kubenswrapper[4846]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 20 10:58:39 crc kubenswrapper[4846]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: continue Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # Append resolver entries for services Mar 20 10:58:39 crc kubenswrapper[4846]: rc=0 Mar 20 10:58:39 crc kubenswrapper[4846]: for svc in "${!svc_ips[@]}"; do Mar 20 10:58:39 crc kubenswrapper[4846]: for ip in ${svc_ips[${svc}]}; do Mar 20 10:58:39 crc kubenswrapper[4846]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ $rc -ne 0 ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: continue Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 20 10:58:39 crc kubenswrapper[4846]: # Replace /etc/hosts with our modified version if needed Mar 20 10:58:39 crc kubenswrapper[4846]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 20 10:58:39 crc kubenswrapper[4846]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 60 & wait Mar 20 10:58:39 crc kubenswrapper[4846]: unset svc_ips Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msz4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-j4h4w_openshift-dns(9275b8c7-ce16-4c4a-a7eb-93d263b751d6): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.813816 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-j4h4w" podUID="9275b8c7-ce16-4c4a-a7eb-93d263b751d6" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.813828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"70aef47ee9ed0f114161f86d0fc5024c5fca18b5c1184bc2120cbb865ee65a82"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.814269 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzg6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.815932 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"1f8b167f2df64ee51562c3a542bd0eb419c5ab206c41a98643f31e50925f9ed3"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.816040 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.816092 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 10:58:39 crc kubenswrapper[4846]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 10:58:39 crc kubenswrapper[4846]: ho_enable="--enable-hybrid-overlay" Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 10:58:39 crc kubenswrapper[4846]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 10:58:39 crc kubenswrapper[4846]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-host=127.0.0.1 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --webhook-port=9743 \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ho_enable} \ Mar 20 10:58:39 crc kubenswrapper[4846]: --enable-interconnect \ Mar 20 10:58:39 crc kubenswrapper[4846]: --disable-approver \ Mar 20 10:58:39 crc kubenswrapper[4846]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --wait-for-kubernetes-api=200s \ Mar 20 10:58:39 crc kubenswrapper[4846]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel="${LOGLEVEL}" Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.816421 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bzg6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.817636 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.817877 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 20 10:58:39 crc kubenswrapper[4846]: apiVersion: v1 Mar 20 10:58:39 crc kubenswrapper[4846]: clusters: Mar 20 10:58:39 crc kubenswrapper[4846]: - cluster: Mar 20 10:58:39 crc kubenswrapper[4846]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 20 10:58:39 crc kubenswrapper[4846]: server: https://api-int.crc.testing:6443 Mar 20 10:58:39 crc kubenswrapper[4846]: name: default-cluster Mar 20 10:58:39 crc kubenswrapper[4846]: contexts: Mar 20 10:58:39 crc kubenswrapper[4846]: - context: Mar 20 10:58:39 crc kubenswrapper[4846]: cluster: default-cluster Mar 20 10:58:39 crc kubenswrapper[4846]: namespace: default Mar 20 10:58:39 crc kubenswrapper[4846]: user: default-auth Mar 20 10:58:39 crc kubenswrapper[4846]: name: default-context Mar 20 10:58:39 crc kubenswrapper[4846]: current-context: default-context Mar 20 10:58:39 crc kubenswrapper[4846]: kind: Config Mar 20 10:58:39 crc kubenswrapper[4846]: preferences: {} Mar 20 10:58:39 crc kubenswrapper[4846]: users: Mar 20 10:58:39 crc kubenswrapper[4846]: - name: default-auth Mar 20 10:58:39 crc kubenswrapper[4846]: user: Mar 20 10:58:39 crc kubenswrapper[4846]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 10:58:39 crc kubenswrapper[4846]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 20 10:58:39 crc kubenswrapper[4846]: EOF Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bv8k6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.818653 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --disable-webhook \ Mar 20 10:58:39 crc kubenswrapper[4846]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel="${LOGLEVEL}" Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.818758 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" event={"ID":"ef1b0cac-65d1-4f15-84f8-a906fa5c2967","Type":"ContainerStarted","Data":"09092057a358bb5fe205456ad6506bec8c696114f0da7ef4f98780b9a18a870a"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.819152 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.819699 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.819755 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 20 10:58:39 crc kubenswrapper[4846]: set -euo pipefail Mar 20 10:58:39 crc kubenswrapper[4846]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 20 10:58:39 crc kubenswrapper[4846]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 20 10:58:39 crc kubenswrapper[4846]: # As the secret mount is optional we must wait for the files to be present. Mar 20 10:58:39 crc kubenswrapper[4846]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 20 10:58:39 crc kubenswrapper[4846]: TS=$(date +%s) Mar 20 10:58:39 crc kubenswrapper[4846]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 20 10:58:39 crc kubenswrapper[4846]: HAS_LOGGED_INFO=0 Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: log_missing_certs(){ Mar 20 10:58:39 crc kubenswrapper[4846]: CUR_TS=$(date +%s) Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 20 10:58:39 crc kubenswrapper[4846]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 20 10:58:39 crc kubenswrapper[4846]: HAS_LOGGED_INFO=1 Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: } Mar 20 10:58:39 crc kubenswrapper[4846]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 20 10:58:39 crc kubenswrapper[4846]: log_missing_certs Mar 20 10:58:39 crc kubenswrapper[4846]: sleep 5 Mar 20 10:58:39 crc kubenswrapper[4846]: done Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/kube-rbac-proxy \ Mar 20 10:58:39 crc kubenswrapper[4846]: --logtostderr \ Mar 20 10:58:39 crc kubenswrapper[4846]: --secure-listen-address=:9108 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 20 10:58:39 crc kubenswrapper[4846]: --upstream=http://127.0.0.1:29108/ \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-private-key-file=${TLS_PK} \ Mar 20 10:58:39 crc kubenswrapper[4846]: --tls-cert-file=${TLS_CERT} Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnzt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-tpdzl_openshift-ovn-kubernetes(ef1b0cac-65d1-4f15-84f8-a906fa5c2967): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.820079 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerStarted","Data":"2a41a2584a16d6113f21eb05326eb10e12fb98db8950aa8e977f81f84194931a"} Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.821367 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 20 10:58:39 crc kubenswrapper[4846]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 20 10:58:39 crc kubenswrapper[4846]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cplvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-cqnfg_openshift-multus(dfe48192-f5b6-4ba5-a4a4-475459fd1815): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.821987 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 10:58:39 crc kubenswrapper[4846]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ -f "/env/_master" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: set -o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: source "/env/_master" Mar 20 10:58:39 crc kubenswrapper[4846]: set +o allexport Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_join_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_join_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_transit_switch_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_transit_switch_subnet_opt= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "" != "" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: dns_name_resolver_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "false" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: persistent_ips_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "true" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: # This is needed so that converting clusters from GA to TP Mar 20 10:58:39 crc kubenswrapper[4846]: # will rollout control plane pods as well Mar 20 10:58:39 crc kubenswrapper[4846]: network_segmentation_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: multi_network_enabled_flag= Mar 20 10:58:39 crc kubenswrapper[4846]: if [[ "true" == "true" ]]; then Mar 20 10:58:39 crc kubenswrapper[4846]: multi_network_enabled_flag="--enable-multi-network" Mar 20 10:58:39 crc kubenswrapper[4846]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 20 10:58:39 crc kubenswrapper[4846]: fi Mar 20 10:58:39 crc kubenswrapper[4846]: Mar 20 10:58:39 crc kubenswrapper[4846]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 20 10:58:39 crc kubenswrapper[4846]: exec /usr/bin/ovnkube \ Mar 20 10:58:39 crc kubenswrapper[4846]: --enable-interconnect \ Mar 20 10:58:39 crc kubenswrapper[4846]: --init-cluster-manager "${K8S_NODE}" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 20 10:58:39 crc kubenswrapper[4846]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-bind-address "127.0.0.1:29108" \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-enable-pprof \ Mar 20 10:58:39 crc kubenswrapper[4846]: --metrics-enable-config-duration \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v4_join_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v6_join_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${dns_name_resolver_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${persistent_ips_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${multi_network_enabled_flag} \ Mar 20 10:58:39 crc kubenswrapper[4846]: ${network_segmentation_enabled_flag} Mar 20 10:58:39 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnzt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-tpdzl_openshift-ovn-kubernetes(ef1b0cac-65d1-4f15-84f8-a906fa5c2967): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 10:58:39 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.823877 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-cqnfg" podUID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.823878 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" podUID="ef1b0cac-65d1-4f15-84f8-a906fa5c2967" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.830769 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.842419 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.858154 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.879890 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.892462 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.903181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.903216 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.903226 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.903245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.903257 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:39Z","lastTransitionTime":"2026-03-20T10:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.906784 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.920179 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.930059 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.945437 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.962698 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.977254 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.994725 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.995002 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:58:40.994967161 +0000 UTC m=+92.361375397 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.995113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.995173 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:39 crc kubenswrapper[4846]: I0320 10:58:39.995007 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.995358 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.995369 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.995451 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:40.995427852 +0000 UTC m=+92.361836118 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:39 crc kubenswrapper[4846]: E0320 10:58:39.995499 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:40.995466453 +0000 UTC m=+92.361874699 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.006472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.006630 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.006716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.006801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.006924 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.031510 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.071406 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.096035 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.096112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.096159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096271 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096391 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:58:41.096362316 +0000 UTC m=+92.462770542 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096398 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096429 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096449 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096519 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:41.096495139 +0000 UTC m=+92.462903575 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096609 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096625 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096638 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.096681 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:41.096667413 +0000 UTC m=+92.463075669 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.111110 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.111163 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.111176 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.111196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.111210 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.113593 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.143653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.143711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.143729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.143753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.143768 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.150358 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.156004 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.161488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.161639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.161666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.161701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.161726 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.175668 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.180082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.180126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.180152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.180173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.180188 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.189608 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.195053 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.199412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.199450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.199523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.199576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.199597 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.215240 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.226575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.226648 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.226658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.226673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.226686 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.232403 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.238319 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.238471 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.240415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.240469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.240485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.240511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.240529 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.269671 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.316872 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.322063 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.322508 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.344617 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.344702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.344721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.344781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.344800 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.354010 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.354429 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.354598 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.378318 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.411221 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.448158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.448202 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.448215 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.448238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.448252 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.455701 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.487756 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.533284 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.551237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.551291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.551306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.551326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.551344 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.568441 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.613956 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.655181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.655265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.655283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.655308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.655328 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.758118 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.758399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.758508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.758574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.758637 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.825335 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:40 crc kubenswrapper[4846]: E0320 10:58:40.825643 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.861579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.861795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.861953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.862120 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.862239 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.965656 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.965730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.965747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.965774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:40 crc kubenswrapper[4846]: I0320 10:58:40.965792 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:40Z","lastTransitionTime":"2026-03-20T10:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.007630 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.007759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.007872 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.007835457 +0000 UTC m=+94.374243723 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.007960 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.008034 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.008013291 +0000 UTC m=+94.374421567 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.008097 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.008223 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.008281 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.008267887 +0000 UTC m=+94.374676163 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.071316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.071928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.072089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.072244 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.072390 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.109736 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.110201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.110397 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110066 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110345 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110814 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110831 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110490 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110891 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110915 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110776 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.110742817 +0000 UTC m=+94.477151083 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110971 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.110948021 +0000 UTC m=+94.477356257 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.110983 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:43.110978632 +0000 UTC m=+94.477386868 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.176628 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.176685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.176701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.176727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.176743 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.280017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.280089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.280113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.280146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.280170 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.322542 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.322562 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.322807 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.322721 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.322954 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:41 crc kubenswrapper[4846]: E0320 10:58:41.323182 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.332056 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.333870 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.336605 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.338128 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.339741 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.340631 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.341488 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.342808 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.343530 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.344675 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.345297 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.346510 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.347036 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.347559 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.348504 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.349085 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.350080 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.350477 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.351096 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.352301 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.352802 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.353874 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.354359 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.355540 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.356137 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.356844 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.358081 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.358595 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.359589 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.360101 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.361062 4846 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.361174 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.363055 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.364173 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.364638 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.366400 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.367248 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.368713 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.369567 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.371147 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.371876 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.373383 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.374285 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.375823 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.376374 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.376980 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.377719 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.378506 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.379002 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.379468 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.379981 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.380498 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.381093 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.381591 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.383555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.383591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.383601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.383623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.383637 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.488063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.488125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.488148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.488182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.488205 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.591594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.591647 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.591658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.591675 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.591686 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.695890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.695991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.696004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.696026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.696042 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.800407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.800469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.800480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.800499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.800512 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.904405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.904490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.904515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.904548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:41 crc kubenswrapper[4846]: I0320 10:58:41.904569 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:41Z","lastTransitionTime":"2026-03-20T10:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.008254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.008333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.008358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.008392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.008412 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.111695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.111759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.111776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.111802 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.111822 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.215301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.215372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.215393 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.215423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.215444 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.318714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.318778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.318797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.318828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.318849 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.322296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:42 crc kubenswrapper[4846]: E0320 10:58:42.322482 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.421985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.422058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.422080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.422128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.422147 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.506833 4846 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.525438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.525495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.525512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.525540 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.525561 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.629281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.629338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.629358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.629388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.629409 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.733619 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.733690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.733708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.733735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.733754 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.836868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.836973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.837025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.837054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.837073 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.941080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.941162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.941183 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.941221 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:42 crc kubenswrapper[4846]: I0320 10:58:42.941245 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:42Z","lastTransitionTime":"2026-03-20T10:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.035271 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.035541 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.035494572 +0000 UTC m=+98.401902819 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.035632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.035719 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.035833 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.036005 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.035968533 +0000 UTC m=+98.402376889 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.036012 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.036093 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.036074986 +0000 UTC m=+98.402483262 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.044402 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.044467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.044486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.044513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.044533 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.137117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.137194 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.137242 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137448 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137477 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137498 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137490 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137570 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137637 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137661 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137585 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.137560223 +0000 UTC m=+98.503968499 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137812 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.137760848 +0000 UTC m=+98.504169244 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.137871 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:47.13784996 +0000 UTC m=+98.504258436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.147932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.147995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.148014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.148039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.148057 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.252258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.252359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.252386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.252423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.252451 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.322498 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.322640 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.322770 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.322875 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.323116 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:43 crc kubenswrapper[4846]: E0320 10:58:43.323783 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.355978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.356055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.356074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.356102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.356123 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.460158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.460234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.460253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.460283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.460305 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.564466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.564606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.564683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.564718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.564739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.669181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.669273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.669292 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.669325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.669348 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.772952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.773006 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.773018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.773036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.773050 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.876156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.876238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.876265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.876306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.876335 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.980668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.980726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.980742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.980767 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:43 crc kubenswrapper[4846]: I0320 10:58:43.980789 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:43Z","lastTransitionTime":"2026-03-20T10:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.084066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.084152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.084174 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.084210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.084233 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.187967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.188041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.188067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.188095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.188115 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.292030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.292113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.292138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.292173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.292198 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.321834 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:44 crc kubenswrapper[4846]: E0320 10:58:44.322203 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.395479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.395541 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.395560 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.395587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.395608 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.498175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.498243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.498260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.498279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.498290 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.602570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.602657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.602679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.602707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.602722 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.704674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.704727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.704740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.704759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.704773 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.808185 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.808265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.808280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.808306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.808322 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.910743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.910791 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.910801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.910817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:44 crc kubenswrapper[4846]: I0320 10:58:44.910830 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:44Z","lastTransitionTime":"2026-03-20T10:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.013757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.013821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.013839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.013868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.013889 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.117555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.117624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.117641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.117662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.117674 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.221061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.221136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.221157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.221188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.221208 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.322624 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.322689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:45 crc kubenswrapper[4846]: E0320 10:58:45.322789 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:45 crc kubenswrapper[4846]: E0320 10:58:45.322990 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.323109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:45 crc kubenswrapper[4846]: E0320 10:58:45.323255 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.325937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.325966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.325977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.325994 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.326008 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.429723 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.429783 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.429803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.429834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.429855 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.533176 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.533246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.533266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.533294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.533313 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.637206 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.637263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.637281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.637306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.637323 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.740316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.740388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.740405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.740449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.740471 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.846751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.846810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.847627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.847704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.847729 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.951696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.951772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.951792 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.951822 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:45 crc kubenswrapper[4846]: I0320 10:58:45.951843 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:45Z","lastTransitionTime":"2026-03-20T10:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.055420 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.055490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.055508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.055536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.055557 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.158557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.158654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.158687 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.158785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.158814 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.262184 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.262257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.262274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.262304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.262324 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.322262 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:46 crc kubenswrapper[4846]: E0320 10:58:46.322498 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.365427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.365472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.365485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.365504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.365517 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.468657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.468704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.468717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.468738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.468751 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.571372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.571448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.571470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.571502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.571529 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.673954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.674029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.674048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.674079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.674099 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.778122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.778193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.778219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.778249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.778271 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.881535 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.881599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.881616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.881641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.881662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.985359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.985430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.985449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.985478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:46 crc kubenswrapper[4846]: I0320 10:58:46.985497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:46Z","lastTransitionTime":"2026-03-20T10:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.084817 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.085159 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.085111201 +0000 UTC m=+106.451519467 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.085251 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.085335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.085480 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.085588 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.085627 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.085587132 +0000 UTC m=+106.451995558 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.085685 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.085662034 +0000 UTC m=+106.452070300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.088069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.088122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.088140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.088166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.088184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.186727 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.186846 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.186964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187135 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187229 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187264 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187268 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187325 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.187268834 +0000 UTC m=+106.553677190 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187331 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187383 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187456 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.187438658 +0000 UTC m=+106.553846934 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187286 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.187543 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:58:55.18752258 +0000 UTC m=+106.553931096 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.192402 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.192462 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.192482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.192513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.192532 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.295779 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.295833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.295851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.295877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.295974 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.322233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.322233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.322495 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.322566 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.322746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:47 crc kubenswrapper[4846]: E0320 10:58:47.322960 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.399561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.400009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.400160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.400299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.400465 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.504323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.504678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.504788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.504949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.505057 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.608832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.608938 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.608955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.608982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.608999 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.712693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.712753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.712771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.712797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.712816 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.815805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.815881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.815931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.815959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.815978 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.919853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.919994 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.920025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.920072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:47 crc kubenswrapper[4846]: I0320 10:58:47.920107 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:47Z","lastTransitionTime":"2026-03-20T10:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.023307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.023378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.023396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.023423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.023449 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.126878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.126981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.126998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.127026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.127040 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.230521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.230573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.230587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.230616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.230641 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.322383 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:48 crc kubenswrapper[4846]: E0320 10:58:48.322628 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.334235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.334310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.334334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.334372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.334396 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.438029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.438109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.438175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.438212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.438235 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.541109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.541186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.541208 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.541235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.541257 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.649038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.649222 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.649238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.649271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.649306 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.752933 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.752997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.753018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.753044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.753063 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.855338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.855970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.856049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.856115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.856178 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.959933 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.959984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.959996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.960019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:48 crc kubenswrapper[4846]: I0320 10:58:48.960032 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:48Z","lastTransitionTime":"2026-03-20T10:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.064035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.064110 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.064138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.064173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.064198 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.168028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.168103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.168125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.168154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.168174 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.272205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.272277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.272294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.272323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.272343 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.322311 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.322424 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:49 crc kubenswrapper[4846]: E0320 10:58:49.322521 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.322362 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:49 crc kubenswrapper[4846]: E0320 10:58:49.322683 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:49 crc kubenswrapper[4846]: E0320 10:58:49.322830 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.334397 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.348439 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.375082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.375212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.375239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.375270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.375324 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.376103 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.389225 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.404077 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.419827 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.432058 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.449032 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.462263 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.477846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.478127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.478235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.478308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.478389 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.481551 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.495727 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.513230 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.529660 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.547584 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.562196 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.581096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.581148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.581160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.581178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.581190 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.684553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.684623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.684641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.684671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.684690 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.788344 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.788425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.788442 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.788466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.788486 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.892019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.892097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.892116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.892147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.892168 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.995712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.995791 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.995816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.995852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:49 crc kubenswrapper[4846]: I0320 10:58:49.995879 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:49Z","lastTransitionTime":"2026-03-20T10:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.100286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.100383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.100407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.100443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.100466 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.203591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.203674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.203693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.203723 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.203743 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.306823 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.306916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.306932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.306962 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.306978 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.321677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.321980 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.386847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.387300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.387471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.387623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.387779 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.405484 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.411150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.411206 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.411225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.411252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.411270 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.427557 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.433258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.433492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.433673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.433871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.434101 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.452952 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.458940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.459013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.459035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.459064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.459086 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.472249 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.477508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.477578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.477594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.477617 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.477633 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.495727 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:50 crc kubenswrapper[4846]: E0320 10:58:50.495990 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.499207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.499619 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.499807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.500055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.500415 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.603349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.603409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.603426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.603454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.603472 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.707035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.707107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.707136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.707170 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.707195 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.810449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.810533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.810560 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.810591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.810615 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.914584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.914688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.914706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.914736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:50 crc kubenswrapper[4846]: I0320 10:58:50.914756 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:50Z","lastTransitionTime":"2026-03-20T10:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.017968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.018036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.018055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.018083 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.018104 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.121665 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.121709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.121729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.121751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.121767 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.225319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.225429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.225452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.225484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.225505 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.328332 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.328369 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.328570 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:51 crc kubenswrapper[4846]: E0320 10:58:51.329407 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:51 crc kubenswrapper[4846]: E0320 10:58:51.329600 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:51 crc kubenswrapper[4846]: E0320 10:58:51.330186 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.331854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.331931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.331958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.331988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.332012 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.436216 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.436281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.436299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.436328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.436347 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.539840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.539955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.540037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.540070 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.540092 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.643366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.643450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.643476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.643511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.643535 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.747040 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.747128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.747154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.747191 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.747216 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.850733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.850795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.850812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.850838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.850857 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.864066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" event={"ID":"ef1b0cac-65d1-4f15-84f8-a906fa5c2967","Type":"ContainerStarted","Data":"9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.864118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" event={"ID":"ef1b0cac-65d1-4f15-84f8-a906fa5c2967","Type":"ContainerStarted","Data":"f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.879807 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.894564 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.916875 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.932234 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.950201 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.955205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.955513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.955683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.955816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.955957 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:51Z","lastTransitionTime":"2026-03-20T10:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.969484 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:51 crc kubenswrapper[4846]: I0320 10:58:51.987674 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.000045 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.012862 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.031068 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.044515 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.058630 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.058695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.058717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.058747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.058768 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.059384 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.072775 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.088722 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.100113 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.161556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.161945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.162134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.162369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.162569 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.266064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.266102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.266139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.266160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.266172 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.322471 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:52 crc kubenswrapper[4846]: E0320 10:58:52.322656 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.322879 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:58:52 crc kubenswrapper[4846]: E0320 10:58:52.323268 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.368968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.369074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.369109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.369148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.369189 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.481565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.481603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.481615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.481634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.481647 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.585805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.585850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.585863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.585887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.585923 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.689259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.689316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.689330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.689361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.689380 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.793317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.793369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.793381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.793400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.793411 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.871540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerStarted","Data":"98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.873864 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7rtn2" event={"ID":"b2799b9e-c7c5-4816-97d6-d181a459f81f","Type":"ContainerStarted","Data":"d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.875663 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerStarted","Data":"3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.877589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j4h4w" event={"ID":"9275b8c7-ce16-4c4a-a7eb-93d263b751d6","Type":"ContainerStarted","Data":"744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.884471 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.884551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.885790 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.888034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.888080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.895562 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.895622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.895636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.895659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.895673 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.900655 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.927681 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.943822 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.955976 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.973354 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.984287 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.998314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.998374 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.998392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.998419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:52 crc kubenswrapper[4846]: I0320 10:58:52.998443 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:52Z","lastTransitionTime":"2026-03-20T10:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.003859 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.036722 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.060812 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.081440 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.092952 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.100921 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.100984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.100998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.101063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.101088 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.108766 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.124073 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.138740 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.154687 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.168559 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.181434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.194450 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.204185 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.204243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.204259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.204289 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.204304 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.206139 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.218252 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.228484 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.264981 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.286321 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.305815 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.306951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.307014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.307033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.307063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.307084 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.321802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.321958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:53 crc kubenswrapper[4846]: E0320 10:58:53.322104 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:53 crc kubenswrapper[4846]: E0320 10:58:53.322539 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.322719 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:53 crc kubenswrapper[4846]: E0320 10:58:53.322819 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.323612 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.341756 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.359093 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.374543 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.392673 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.409858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.409933 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.409943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.409962 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.409972 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.513484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.514034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.514045 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.514064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.514077 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.617482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.617536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.617549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.617569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.617584 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.720737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.720777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.720786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.720800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.720810 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.824480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.824536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.824546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.824566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.824576 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.894248 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274" exitCode=0 Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.894315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.912619 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.925384 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.927270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.927312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.927325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.927365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.927381 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:53Z","lastTransitionTime":"2026-03-20T10:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.942371 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.966706 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.982425 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:53 crc kubenswrapper[4846]: I0320 10:58:53.998061 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.013930 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.027672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.030449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.030495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.030509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.030530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.030546 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.050864 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.071687 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.087990 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.103082 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.116831 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.131295 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.133237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.133301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.133310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.133330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.133342 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.144780 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.236565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.236657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.236681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.236714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.236736 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.321883 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:54 crc kubenswrapper[4846]: E0320 10:58:54.322191 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.339081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.339125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.339135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.339171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.339183 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.442922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.442980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.442997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.443022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.443052 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.545658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.545729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.545744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.545768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.545783 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.648737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.648789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.648798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.648819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.648830 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.752258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.752300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.752311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.752332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.752344 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.857188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.857795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.857822 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.857861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.857888 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.901339 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa" exitCode=0 Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.901434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.904167 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.909151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.923269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.938748 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.953817 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.962812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.962975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.963007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.963044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.963067 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:54Z","lastTransitionTime":"2026-03-20T10:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.971320 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:54 crc kubenswrapper[4846]: I0320 10:58:54.988319 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.016570 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.039777 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.055181 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.066189 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.066244 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.066258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.066288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.066311 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.067824 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.083142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.097535 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.100472 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.100617 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.100596804 +0000 UTC m=+122.467005040 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.100689 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.100724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.100786 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.100832 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.100841 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.100832849 +0000 UTC m=+122.467241085 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.100857 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.10085067 +0000 UTC m=+122.467258906 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.110010 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.123207 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.138834 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.149893 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.167124 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.168356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.168412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.168429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.168455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.168469 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.182434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.195674 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.202277 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.202360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.202427 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.202753 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.202797 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.202824 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.202945 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.20288103 +0000 UTC m=+122.569289306 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203392 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203452 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.203434042 +0000 UTC m=+122.569842318 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203537 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203555 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203571 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.203610 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:11.203597486 +0000 UTC m=+122.570005762 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.204468 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.222601 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.236758 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.252637 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.264888 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.276519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.276583 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.276601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.276627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.276646 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.286288 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.304710 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.316125 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.322540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.322548 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.322635 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.322699 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.322866 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:55 crc kubenswrapper[4846]: E0320 10:58:55.323006 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.338039 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.357750 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.376331 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.380698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.380750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.380768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.380798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.380816 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.394580 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.484538 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.484616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.484633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.484659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.484675 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.587528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.587602 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.587614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.587638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.587650 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.690621 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.690999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.691106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.691358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.691455 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.794674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.794715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.794727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.794745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.794756 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.897549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.897599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.897608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.897627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.897640 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:55Z","lastTransitionTime":"2026-03-20T10:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.915159 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2" exitCode=0 Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.915253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.916747 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" exitCode=0 Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.916804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.940921 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.960128 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:55 crc kubenswrapper[4846]: I0320 10:58:55.982802 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.000840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.000957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.000976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.001407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.001436 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.002339 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.025024 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.043057 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.058260 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.073073 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.085517 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.100448 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.106185 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.106223 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.106235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.106274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.106288 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.113468 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.130224 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.143480 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.170792 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.201513 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.209067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.209100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.209109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.209137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.209147 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.230582 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.260439 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.276025 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.292553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.305438 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.311310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.311343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.311352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.311369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.311379 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.321883 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:56 crc kubenswrapper[4846]: E0320 10:58:56.322050 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.331116 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.346034 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.360731 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.378286 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.397841 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.414291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.414348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.414361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.414383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.414398 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.416568 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.430218 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.440642 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.459175 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.470464 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.516955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.517013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.517032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.517059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.517074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.620318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.620378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.620389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.620412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.620427 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.723610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.723691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.723712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.723747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.723781 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.827655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.827716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.827734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.827763 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.827780 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.927411 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de" exitCode=0 Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.927504 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.932169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.932220 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.932237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.932268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.932293 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:56Z","lastTransitionTime":"2026-03-20T10:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941646 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941685 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941699 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.941715 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.951019 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.972444 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:56 crc kubenswrapper[4846]: I0320 10:58:56.987670 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:56Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.006192 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.020858 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.040655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.040726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.040745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.040776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.040796 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.053234 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.069338 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.083246 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.102721 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.119566 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.142313 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.144012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.144062 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.144072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.144187 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.144205 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.164112 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.177665 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.191286 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.202999 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.246634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.246676 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.246688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.246709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.246721 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.322242 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.322329 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:57 crc kubenswrapper[4846]: E0320 10:58:57.322501 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.323163 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:57 crc kubenswrapper[4846]: E0320 10:58:57.323334 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:57 crc kubenswrapper[4846]: E0320 10:58:57.323542 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.349502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.349553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.349567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.349589 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.349603 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.452851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.452959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.452978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.453008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.453049 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.556877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.556994 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.557018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.557053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.557076 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.660162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.660274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.660294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.660322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.660341 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.764390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.764467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.764484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.764515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.764534 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.867997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.868074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.868098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.868131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.868156 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.949417 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007" exitCode=0 Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.949499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.971304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.971400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.971423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.971480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.971506 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:57Z","lastTransitionTime":"2026-03-20T10:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.976379 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:57 crc kubenswrapper[4846]: I0320 10:58:57.996257 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:57Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.018993 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.040296 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.059322 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.075706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.075803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.075825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.075859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.075889 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.092303 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.117376 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.141281 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.163788 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.178709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.178790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.178803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.178824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.178838 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.181398 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.198547 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.221128 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.236861 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.263576 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.282537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.282604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.282621 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.282644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.282658 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.288203 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.322221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:58:58 crc kubenswrapper[4846]: E0320 10:58:58.322445 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.385555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.385614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.385627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.385647 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.385662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.489089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.489162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.489182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.489211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.489229 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.592809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.593331 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.593346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.593366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.593377 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.697364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.697610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.697630 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.697657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.697677 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.801214 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.801274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.801291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.801317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.801335 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.905019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.905122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.905143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.905506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.905542 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:58Z","lastTransitionTime":"2026-03-20T10:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.958069 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8533b04-cc98-4712-95b3-1dee5598b9b3" containerID="3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f" exitCode=0 Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.958156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerDied","Data":"3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.969945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 10:58:58 crc kubenswrapper[4846]: I0320 10:58:58.986691 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:58Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008166 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.008877 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.040976 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.067340 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.086276 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.103263 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.112736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.112821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.112831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.112851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.112864 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.117291 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.134754 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.152112 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.173967 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.189853 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.205745 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.215211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.215271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.215288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.215317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.215336 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.222247 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.237308 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.249684 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.318768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.318832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.318845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.318874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.318887 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.321760 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.321789 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.321850 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:58:59 crc kubenswrapper[4846]: E0320 10:58:59.321932 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:58:59 crc kubenswrapper[4846]: E0320 10:58:59.322086 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:58:59 crc kubenswrapper[4846]: E0320 10:58:59.322215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.338746 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.361018 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.379767 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.406043 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422419 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.422827 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.435274 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.449968 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.473693 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.507059 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.525408 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.526150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.526235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.526252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.526276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.526295 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.541917 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.558339 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.573610 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.596572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.616834 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.629746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.629802 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.629815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.629835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.629847 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.733689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.733758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.733777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.733806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.733825 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.836610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.836679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.836693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.836716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.836730 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.939858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.939949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.939969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.939995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.940013 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:58:59Z","lastTransitionTime":"2026-03-20T10:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.980753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" event={"ID":"d8533b04-cc98-4712-95b3-1dee5598b9b3","Type":"ContainerStarted","Data":"6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14"} Mar 20 10:58:59 crc kubenswrapper[4846]: I0320 10:58:59.998698 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:58:59Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.015600 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.035988 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.043400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.043479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.043496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.043554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.043575 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.054599 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.070415 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.085741 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.109554 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.130401 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.147008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.147309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.147592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.147910 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.148139 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.152960 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.169405 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.180088 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.194577 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.208570 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.223309 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.236523 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.250355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.250390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.250403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.250421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.250432 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.321653 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.321852 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.353469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.353509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.353519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.353535 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.353544 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.457043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.457108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.457129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.457156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.457176 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.537202 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.537279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.537299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.537325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.537345 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.558248 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.565096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.565471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.565753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.566027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.566229 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.587564 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.593841 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.594125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.594165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.594190 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.594204 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.618803 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.623683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.623723 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.623732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.623748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.623762 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.642241 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.647743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.647798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.647816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.647840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.647859 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.665993 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:00 crc kubenswrapper[4846]: E0320 10:59:00.666276 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.669339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.669582 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.669601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.669633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.669655 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.773711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.773777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.773796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.773824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.773844 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.877115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.877165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.877178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.877198 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.877211 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.979315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.979379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.979392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.979413 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:00 crc kubenswrapper[4846]: I0320 10:59:00.979454 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:00Z","lastTransitionTime":"2026-03-20T10:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.083233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.083288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.083311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.083340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.083359 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.186514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.186576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.186592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.186618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.186636 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.290272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.290326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.290340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.290433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.290448 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.322284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.322397 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.322284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:01 crc kubenswrapper[4846]: E0320 10:59:01.322556 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:01 crc kubenswrapper[4846]: E0320 10:59:01.322837 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:01 crc kubenswrapper[4846]: E0320 10:59:01.323034 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.393982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.394039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.394058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.394089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.394109 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.497196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.497260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.497280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.497308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.497326 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.601166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.601228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.601240 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.601265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.601277 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.711345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.711400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.711414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.711434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.711446 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.815075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.815131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.815150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.815172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.815192 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.918004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.918072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.918090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.918115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.918130 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:01Z","lastTransitionTime":"2026-03-20T10:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.994806 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7"} Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.995165 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.995287 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:01 crc kubenswrapper[4846]: I0320 10:59:01.995439 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.013630 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.023712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.023769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.023787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.023817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.023836 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.031394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.031533 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.032693 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.059409 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.097379 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.131073 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.132919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.132969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.132981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.132999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.133013 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.144823 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.153484 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.177618 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.195070 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.211562 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.222657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.236721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.237624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.237662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.237692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.237715 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.238822 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.252958 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.266415 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.277572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.291770 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.303201 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.317673 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.321535 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:02 crc kubenswrapper[4846]: E0320 10:59:02.321789 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.331185 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.340713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.340750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.340759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.340776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.340788 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.343930 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.364188 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.379044 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.390347 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.404656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.417443 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.429404 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.443770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.443849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.443863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.443886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.443930 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.450683 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.464603 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.484176 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.497934 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.546167 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.546218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.546230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.546249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.546261 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.649250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.649337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.649359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.649388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.649410 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.753552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.753644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.753685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.753713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.753730 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.856758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.856815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.856828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.856848 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.856861 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.959816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.959879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.959910 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.959930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:02 crc kubenswrapper[4846]: I0320 10:59:02.959944 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:02Z","lastTransitionTime":"2026-03-20T10:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.063063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.063119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.063135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.063151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.063162 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.166006 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.166052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.166064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.166081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.166093 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.269090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.269142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.269152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.269172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.269184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.322099 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.322162 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.322212 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:03 crc kubenswrapper[4846]: E0320 10:59:03.322294 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:03 crc kubenswrapper[4846]: E0320 10:59:03.322481 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:03 crc kubenswrapper[4846]: E0320 10:59:03.322628 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.323547 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.342011 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.372559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.372634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.372650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.372696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.372712 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.475669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.475743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.475757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.475775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.475787 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.578881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.578946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.578956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.578972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.578982 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.681184 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.681231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.681243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.681260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.681273 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.788873 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.788984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.789003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.789034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.789056 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.891809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.891855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.891868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.891888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.891926 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.996483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.996581 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.996611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.996645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:03 crc kubenswrapper[4846]: I0320 10:59:03.996665 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:03Z","lastTransitionTime":"2026-03-20T10:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.006507 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.009676 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.010609 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.030211 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.045714 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.061650 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.076667 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.089998 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.098883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.098939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.098951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.098975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.098989 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.102622 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.119820 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.135582 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.157941 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.181562 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.195824 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.201399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.201463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.201476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.201494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.201512 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.216182 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.230039 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.246920 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.267881 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.283109 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.304801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.304856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.304866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.304882 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.304912 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.321425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:04 crc kubenswrapper[4846]: E0320 10:59:04.321560 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.407444 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.407498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.407511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.407529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.407542 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.510713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.510799 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.510818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.510844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.510862 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.614432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.614486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.614496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.614513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.614525 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.718771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.718883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.718977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.719019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.719047 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.821482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.821565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.821576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.821592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.821605 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.923882 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.923941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.923951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.923969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:04 crc kubenswrapper[4846]: I0320 10:59:04.923982 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:04Z","lastTransitionTime":"2026-03-20T10:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.016108 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/0.log" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.020411 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7" exitCode=1 Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.020489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.022238 4846 scope.go:117] "RemoveContainer" containerID="ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.025676 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.025729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.025739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.025758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.025771 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.050192 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.069969 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.090164 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.109663 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.124291 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.130098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.130131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.130140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.130157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.130167 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.142825 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.156440 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.171081 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.186567 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.201834 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.215127 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.228804 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.233148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.233203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.233216 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.233245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.233260 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.242610 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.267395 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:04Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 3 for removal\\\\nI0320 10:59:03.905863 6748 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 10:59:03.905881 6748 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 10:59:03.905890 6748 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 10:59:03.905915 6748 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 10:59:03.905929 6748 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 10:59:03.906030 6748 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906199 6748 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906247 6748 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906368 6748 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906564 6748 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.908020 6748 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 10:59:03.908067 6748 factory.go:656] Stopping watch factory\\\\nI0320 10:59:03.908087 6748 ovnkube.go:599] Stopped ovnkube\\\\nI0320 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.284716 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.300130 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.321944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.321973 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.322037 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:05 crc kubenswrapper[4846]: E0320 10:59:05.322096 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:05 crc kubenswrapper[4846]: E0320 10:59:05.322249 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:05 crc kubenswrapper[4846]: E0320 10:59:05.322367 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.335304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.335351 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.335363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.335392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.335405 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.438152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.438209 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.438218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.438233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.438247 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.541950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.541997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.542008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.542026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.542038 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.645322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.645358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.645369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.645383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.645392 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.748140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.748186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.748198 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.748221 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.748236 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.851034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.851085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.851098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.851114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.851127 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.954395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.954481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.954509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.954542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:05 crc kubenswrapper[4846]: I0320 10:59:05.954573 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:05Z","lastTransitionTime":"2026-03-20T10:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.026783 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/0.log" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.029322 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.030145 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.046733 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.057837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.057911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.057928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.057948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.057960 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.066471 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.082806 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.101472 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.112864 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.128615 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.140597 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.152006 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.160104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.160134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.160143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.160158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.160169 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.163951 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.177656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.204921 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:04Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 3 for removal\\\\nI0320 10:59:03.905863 6748 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 10:59:03.905881 6748 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 10:59:03.905890 6748 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 10:59:03.905915 6748 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 10:59:03.905929 6748 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 10:59:03.906030 6748 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906199 6748 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906247 6748 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906368 6748 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906564 6748 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.908020 6748 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 10:59:03.908067 6748 factory.go:656] Stopping watch factory\\\\nI0320 10:59:03.908087 6748 ovnkube.go:599] Stopped ovnkube\\\\nI0320 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.222274 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.236868 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.257852 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.264590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.264629 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.264641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.264662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.264676 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.280922 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.293473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:06Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.321985 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:06 crc kubenswrapper[4846]: E0320 10:59:06.322290 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.367838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.367915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.367929 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.367956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.367973 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.471035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.471106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.471127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.471162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.471184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.574622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.574699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.574720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.574751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.574772 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.678320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.678389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.678415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.678447 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.678470 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.780946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.781060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.781082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.781111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.781133 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.884533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.884646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.884662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.884685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.884704 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.988998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.989100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.989119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.989153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:06 crc kubenswrapper[4846]: I0320 10:59:06.989178 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:06Z","lastTransitionTime":"2026-03-20T10:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.037343 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/1.log" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.038708 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/0.log" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.044991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.045143 4846 scope.go:117] "RemoveContainer" containerID="ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.044788 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c" exitCode=1 Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.046220 4846 scope.go:117] "RemoveContainer" containerID="cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c" Mar 20 10:59:07 crc kubenswrapper[4846]: E0320 10:59:07.046506 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.070075 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.090565 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.092967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.093024 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.093038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.093060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.093078 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.110100 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.128605 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.146139 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.160680 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.177147 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.195126 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.196358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.196435 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.196458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.196493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.196516 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.210446 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.223736 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.236459 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.268714 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea581fbec98d6e9e432994446a56a477243d66fcc582995ea2347273aee417b7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:04Z\\\",\\\"message\\\":\\\"ending *v1.Pod event handler 3 for removal\\\\nI0320 10:59:03.905863 6748 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 10:59:03.905881 6748 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 10:59:03.905890 6748 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 10:59:03.905915 6748 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 10:59:03.905929 6748 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 10:59:03.906030 6748 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906199 6748 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906247 6748 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906368 6748 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.906564 6748 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0320 10:59:03.908020 6748 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 10:59:03.908067 6748 factory.go:656] Stopping watch factory\\\\nI0320 10:59:03.908087 6748 ovnkube.go:599] Stopped ovnkube\\\\nI0320 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.288271 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.300989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.301071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.301093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.301124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.301145 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.304710 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.322148 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.322211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:07 crc kubenswrapper[4846]: E0320 10:59:07.322562 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.322253 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:07 crc kubenswrapper[4846]: E0320 10:59:07.322605 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:07 crc kubenswrapper[4846]: E0320 10:59:07.322711 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.327524 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.346452 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:07Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.404258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.404578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.404685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.404815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.404922 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.509421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.509489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.509509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.509537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.509560 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.613058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.613116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.613129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.613151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.613164 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.716055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.716363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.716470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.716616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.716718 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.820726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.821126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.821278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.821433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.821616 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.924845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.924951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.924974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.925007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:07 crc kubenswrapper[4846]: I0320 10:59:07.925029 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:07Z","lastTransitionTime":"2026-03-20T10:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.029115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.029172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.029186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.029210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.029229 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.050796 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/1.log" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.055885 4846 scope.go:117] "RemoveContainer" containerID="cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c" Mar 20 10:59:08 crc kubenswrapper[4846]: E0320 10:59:08.056203 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.067642 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.079260 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.091840 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.123088 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.132618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.132937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.133380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.133502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.133608 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.144064 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.159016 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.174876 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.190203 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.202712 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.223852 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.239267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.239516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.239620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.239709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.239785 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.241121 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.259482 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.279550 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.298412 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.312713 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.321718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:08 crc kubenswrapper[4846]: E0320 10:59:08.321870 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.327401 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:08Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.342521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.342678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.342810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.342967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.343084 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.446360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.446446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.446472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.446499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.446517 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.549154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.549218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.549242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.549270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.549286 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.652755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.652812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.652828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.652853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.652870 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.756429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.756791 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.756874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.756977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.757044 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.859856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.859981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.860001 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.860062 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.860083 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.964757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.964841 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.964865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.964929 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:08 crc kubenswrapper[4846]: I0320 10:59:08.964950 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:08Z","lastTransitionTime":"2026-03-20T10:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.068765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.068874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.068892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.068958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.068977 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:09Z","lastTransitionTime":"2026-03-20T10:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.172354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.172448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.172481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.172519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.172542 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:09Z","lastTransitionTime":"2026-03-20T10:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:09 crc kubenswrapper[4846]: E0320 10:59:09.273516 4846 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.333323 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.333373 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.333398 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:09 crc kubenswrapper[4846]: E0320 10:59:09.333523 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:09 crc kubenswrapper[4846]: E0320 10:59:09.333943 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:09 crc kubenswrapper[4846]: E0320 10:59:09.334098 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.351657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.372078 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.386562 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.399339 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.412214 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.425401 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.457924 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.484090 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: E0320 10:59:09.494027 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.499426 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.523949 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.542800 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.560722 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.576364 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.588763 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.602096 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:09 crc kubenswrapper[4846]: I0320 10:59:09.616512 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:09Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.322201 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.322455 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.860868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.860936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.860968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.860985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.860997 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:10Z","lastTransitionTime":"2026-03-20T10:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.878170 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.882342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.882382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.882395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.882412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.882425 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:10Z","lastTransitionTime":"2026-03-20T10:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.896332 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.902076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.902154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.902173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.902201 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.902221 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:10Z","lastTransitionTime":"2026-03-20T10:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.918034 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.923345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.923408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.923421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.923441 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.923484 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:10Z","lastTransitionTime":"2026-03-20T10:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.942417 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.948125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.948192 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.948207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.948235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:10 crc kubenswrapper[4846]: I0320 10:59:10.948256 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:10Z","lastTransitionTime":"2026-03-20T10:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.963283 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:10 crc kubenswrapper[4846]: E0320 10:59:10.963510 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.108163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.108209 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.108184317 +0000 UTC m=+154.474592553 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.108331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.108367 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.108479 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.108488 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.108541 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.108531375 +0000 UTC m=+154.474939611 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.108562 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.108554655 +0000 UTC m=+154.474962891 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.209921 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.209972 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.210000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210100 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210131 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210149 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210162 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210174 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.210156706 +0000 UTC m=+154.576564942 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210195 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.210183646 +0000 UTC m=+154.576591882 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210217 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210265 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210280 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.210351 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 10:59:43.21033398 +0000 UTC m=+154.576742216 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.322424 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.322490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:11 crc kubenswrapper[4846]: I0320 10:59:11.322424 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.322673 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.322822 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:11 crc kubenswrapper[4846]: E0320 10:59:11.323013 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:12 crc kubenswrapper[4846]: I0320 10:59:12.322243 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:12 crc kubenswrapper[4846]: E0320 10:59:12.322550 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:13 crc kubenswrapper[4846]: I0320 10:59:13.321966 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:13 crc kubenswrapper[4846]: I0320 10:59:13.321977 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:13 crc kubenswrapper[4846]: I0320 10:59:13.322174 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:13 crc kubenswrapper[4846]: E0320 10:59:13.322255 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:13 crc kubenswrapper[4846]: E0320 10:59:13.322507 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:13 crc kubenswrapper[4846]: E0320 10:59:13.322405 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:14 crc kubenswrapper[4846]: I0320 10:59:14.322008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:14 crc kubenswrapper[4846]: E0320 10:59:14.322233 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:14 crc kubenswrapper[4846]: E0320 10:59:14.495798 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:15 crc kubenswrapper[4846]: I0320 10:59:15.322243 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:15 crc kubenswrapper[4846]: I0320 10:59:15.322315 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:15 crc kubenswrapper[4846]: E0320 10:59:15.322429 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:15 crc kubenswrapper[4846]: E0320 10:59:15.322620 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:15 crc kubenswrapper[4846]: I0320 10:59:15.322261 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:15 crc kubenswrapper[4846]: E0320 10:59:15.322746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:16 crc kubenswrapper[4846]: I0320 10:59:16.321944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:16 crc kubenswrapper[4846]: E0320 10:59:16.322195 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:17 crc kubenswrapper[4846]: I0320 10:59:17.321538 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:17 crc kubenswrapper[4846]: I0320 10:59:17.321681 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:17 crc kubenswrapper[4846]: E0320 10:59:17.321849 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:17 crc kubenswrapper[4846]: I0320 10:59:17.321960 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:17 crc kubenswrapper[4846]: E0320 10:59:17.322181 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:17 crc kubenswrapper[4846]: E0320 10:59:17.322303 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:18 crc kubenswrapper[4846]: I0320 10:59:18.321722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:18 crc kubenswrapper[4846]: E0320 10:59:18.321982 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.321655 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.321701 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:19 crc kubenswrapper[4846]: E0320 10:59:19.321947 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.322048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:19 crc kubenswrapper[4846]: E0320 10:59:19.322292 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:19 crc kubenswrapper[4846]: E0320 10:59:19.322396 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.383133 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.400152 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.421331 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.445292 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.461873 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.480766 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: E0320 10:59:19.497978 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.501026 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.523497 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.551285 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.570812 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.589104 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.606148 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.640182 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.663108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.692299 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:19 crc kubenswrapper[4846]: I0320 10:59:19.712795 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:19Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:20 crc kubenswrapper[4846]: I0320 10:59:20.322259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:20 crc kubenswrapper[4846]: E0320 10:59:20.322537 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.369532 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.369551 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.369708 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.369841 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.370478 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.370570 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.371206 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.371276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.371297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.371322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.371340 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:21Z","lastTransitionTime":"2026-03-20T10:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.386459 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.392235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.392288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.392308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.392333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.392350 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:21Z","lastTransitionTime":"2026-03-20T10:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.413579 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.419501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.419564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.419583 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.419611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.419631 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:21Z","lastTransitionTime":"2026-03-20T10:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.441580 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.447001 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.447078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.447111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.447147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.447168 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:21Z","lastTransitionTime":"2026-03-20T10:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.469946 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.475839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.476004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.476027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.476056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.476078 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:21Z","lastTransitionTime":"2026-03-20T10:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.493002 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: E0320 10:59:21.493249 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.696423 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.717004 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.737410 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.758574 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.781326 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.803469 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.820847 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.837715 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.853717 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.886371 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.907580 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.924481 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.940425 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.953545 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.969042 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:21 crc kubenswrapper[4846]: I0320 10:59:21.986170 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:21Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:22 crc kubenswrapper[4846]: I0320 10:59:22.006553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:22Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:22 crc kubenswrapper[4846]: I0320 10:59:22.322454 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:22 crc kubenswrapper[4846]: E0320 10:59:22.323283 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:22 crc kubenswrapper[4846]: I0320 10:59:22.323821 4846 scope.go:117] "RemoveContainer" containerID="cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.322259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.322372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:23 crc kubenswrapper[4846]: E0320 10:59:23.322866 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.322279 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:23 crc kubenswrapper[4846]: E0320 10:59:23.323088 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:23 crc kubenswrapper[4846]: E0320 10:59:23.323168 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.340412 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.386287 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/1.log" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.390244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766"} Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.412634 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.426799 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.460208 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.480819 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.500419 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.521224 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.533402 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.555201 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.567228 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.583232 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.597838 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.620866 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.639694 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.655857 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.672657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.697491 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:23 crc kubenswrapper[4846]: I0320 10:59:23.720683 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:23Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.322535 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:24 crc kubenswrapper[4846]: E0320 10:59:24.322776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.398603 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/2.log" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.399453 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/1.log" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.403396 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" exitCode=1 Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.403466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766"} Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.403597 4846 scope.go:117] "RemoveContainer" containerID="cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.404410 4846 scope.go:117] "RemoveContainer" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" Mar 20 10:59:24 crc kubenswrapper[4846]: E0320 10:59:24.404641 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.425688 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.448656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.464533 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.482873 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: E0320 10:59:24.500336 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.504023 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.518262 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.530395 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.550103 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.564218 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.598612 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.620338 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.636642 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.652324 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.672699 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.686578 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.701725 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:24 crc kubenswrapper[4846]: I0320 10:59:24.719081 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:25 crc kubenswrapper[4846]: I0320 10:59:25.322704 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:25 crc kubenswrapper[4846]: I0320 10:59:25.323266 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:25 crc kubenswrapper[4846]: I0320 10:59:25.323193 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:25 crc kubenswrapper[4846]: E0320 10:59:25.323473 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:25 crc kubenswrapper[4846]: E0320 10:59:25.323617 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:25 crc kubenswrapper[4846]: E0320 10:59:25.323737 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:25 crc kubenswrapper[4846]: I0320 10:59:25.409566 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/2.log" Mar 20 10:59:26 crc kubenswrapper[4846]: I0320 10:59:26.322430 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:26 crc kubenswrapper[4846]: E0320 10:59:26.322645 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:27 crc kubenswrapper[4846]: I0320 10:59:27.322064 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:27 crc kubenswrapper[4846]: I0320 10:59:27.322138 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:27 crc kubenswrapper[4846]: I0320 10:59:27.322160 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:27 crc kubenswrapper[4846]: E0320 10:59:27.322301 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:27 crc kubenswrapper[4846]: E0320 10:59:27.322468 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:27 crc kubenswrapper[4846]: E0320 10:59:27.322721 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:28 crc kubenswrapper[4846]: I0320 10:59:28.321960 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:28 crc kubenswrapper[4846]: E0320 10:59:28.322139 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.322508 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.322557 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:29 crc kubenswrapper[4846]: E0320 10:59:29.322762 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.322840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:29 crc kubenswrapper[4846]: E0320 10:59:29.323134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:29 crc kubenswrapper[4846]: E0320 10:59:29.323203 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.359154 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cdd5955115c2ba3c52d87d49cb7a78a04a7b98be0a68ca3e48bbabfba378f78c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:06Z\\\",\\\"message\\\":\\\"lse, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 10:59:06.240949 6889 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nF0320 10:59:06.240971 6889 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.381672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.435154 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.452079 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.466073 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.479362 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.493339 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: E0320 10:59:29.502234 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.523471 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.540069 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.556533 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.573086 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.589745 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.608533 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.628768 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.646991 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.669953 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:29 crc kubenswrapper[4846]: I0320 10:59:29.687416 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:30 crc kubenswrapper[4846]: I0320 10:59:30.322601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:30 crc kubenswrapper[4846]: E0320 10:59:30.323078 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:30 crc kubenswrapper[4846]: I0320 10:59:30.336071 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.322428 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.322428 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.322582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.322808 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.323105 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.323342 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.823510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.823575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.823592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.823618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.823635 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:31Z","lastTransitionTime":"2026-03-20T10:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.845595 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:31Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.852523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.853359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.853586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.853804 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.854036 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:31Z","lastTransitionTime":"2026-03-20T10:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.877020 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:31Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.882870 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.882993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.883013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.883040 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.883060 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:31Z","lastTransitionTime":"2026-03-20T10:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.905715 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:31Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.912205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.912275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.912302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.912333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.912352 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:31Z","lastTransitionTime":"2026-03-20T10:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.935750 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:31Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.941270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.941335 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.941359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.941394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:31 crc kubenswrapper[4846]: I0320 10:59:31.941417 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:31Z","lastTransitionTime":"2026-03-20T10:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.963420 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:31Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:31 crc kubenswrapper[4846]: E0320 10:59:31.963887 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:32 crc kubenswrapper[4846]: I0320 10:59:32.322718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:32 crc kubenswrapper[4846]: E0320 10:59:32.323635 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:33 crc kubenswrapper[4846]: I0320 10:59:33.322337 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:33 crc kubenswrapper[4846]: E0320 10:59:33.322517 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:33 crc kubenswrapper[4846]: I0320 10:59:33.322649 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:33 crc kubenswrapper[4846]: I0320 10:59:33.322337 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:33 crc kubenswrapper[4846]: E0320 10:59:33.322876 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:33 crc kubenswrapper[4846]: E0320 10:59:33.323225 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:34 crc kubenswrapper[4846]: I0320 10:59:34.321740 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:34 crc kubenswrapper[4846]: E0320 10:59:34.322016 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:34 crc kubenswrapper[4846]: E0320 10:59:34.504229 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:35 crc kubenswrapper[4846]: I0320 10:59:35.322307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:35 crc kubenswrapper[4846]: I0320 10:59:35.322363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:35 crc kubenswrapper[4846]: E0320 10:59:35.322518 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:35 crc kubenswrapper[4846]: I0320 10:59:35.322572 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:35 crc kubenswrapper[4846]: E0320 10:59:35.322731 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:35 crc kubenswrapper[4846]: E0320 10:59:35.322880 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.322315 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:36 crc kubenswrapper[4846]: E0320 10:59:36.323475 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.323603 4846 scope.go:117] "RemoveContainer" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" Mar 20 10:59:36 crc kubenswrapper[4846]: E0320 10:59:36.324144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.342404 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.364087 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.382618 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.400067 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.418222 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.431111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.457315 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.482131 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.506031 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.527459 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.553198 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.571739 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.594532 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.614553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.635141 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.655525 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.680710 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:36 crc kubenswrapper[4846]: I0320 10:59:36.694402 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:37 crc kubenswrapper[4846]: I0320 10:59:37.323302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:37 crc kubenswrapper[4846]: I0320 10:59:37.323384 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:37 crc kubenswrapper[4846]: E0320 10:59:37.323538 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:37 crc kubenswrapper[4846]: I0320 10:59:37.323621 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:37 crc kubenswrapper[4846]: E0320 10:59:37.323786 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:37 crc kubenswrapper[4846]: E0320 10:59:37.323942 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:38 crc kubenswrapper[4846]: I0320 10:59:38.322019 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:38 crc kubenswrapper[4846]: E0320 10:59:38.322272 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.321751 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.321846 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.321884 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:39 crc kubenswrapper[4846]: E0320 10:59:39.322034 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:39 crc kubenswrapper[4846]: E0320 10:59:39.322150 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:39 crc kubenswrapper[4846]: E0320 10:59:39.322252 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.337465 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.350589 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.363522 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.395569 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.411178 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.423047 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.435242 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.467164 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.489386 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: E0320 10:59:39.505067 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.517982 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.534950 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.551157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.564476 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.585328 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.598099 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.608497 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.623287 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.636812 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:39Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.698474 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:39 crc kubenswrapper[4846]: I0320 10:59:39.699930 4846 scope.go:117] "RemoveContainer" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" Mar 20 10:59:39 crc kubenswrapper[4846]: E0320 10:59:39.700310 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.322604 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:40 crc kubenswrapper[4846]: E0320 10:59:40.324153 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.473865 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/0.log" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.473956 4846 generic.go:334] "Generic (PLEG): container finished" podID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" containerID="98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f" exitCode=1 Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.473996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerDied","Data":"98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f"} Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.474452 4846 scope.go:117] "RemoveContainer" containerID="98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.510648 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.532604 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.553463 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.570757 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.588364 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.603196 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.617047 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.633271 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.646971 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.660331 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.673915 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.684526 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.705553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.721673 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.738088 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.754200 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.770843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:40 crc kubenswrapper[4846]: I0320 10:59:40.786310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.322431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:41 crc kubenswrapper[4846]: E0320 10:59:41.322666 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.323091 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:41 crc kubenswrapper[4846]: E0320 10:59:41.323199 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.323434 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:41 crc kubenswrapper[4846]: E0320 10:59:41.323526 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.481926 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/0.log" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.482035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerStarted","Data":"8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9"} Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.507664 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.527436 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.547548 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.569720 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.590880 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.609863 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.623224 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.645868 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.664214 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.680357 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.695596 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.729943 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.753787 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.776817 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.800783 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.819168 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.841337 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:41 crc kubenswrapper[4846]: I0320 10:59:41.859509 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:41Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.182828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.182942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.182966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.182995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.183014 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:42Z","lastTransitionTime":"2026-03-20T10:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.207772 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.213712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.213792 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.213812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.213845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.213867 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:42Z","lastTransitionTime":"2026-03-20T10:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.235961 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.242387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.242463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.242486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.242520 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.242539 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:42Z","lastTransitionTime":"2026-03-20T10:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.267435 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.273784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.274057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.274076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.274108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.274135 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:42Z","lastTransitionTime":"2026-03-20T10:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.294366 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.299176 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.299225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.299243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.299267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.299289 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:42Z","lastTransitionTime":"2026-03-20T10:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.319780 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:42Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.320042 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:42 crc kubenswrapper[4846]: I0320 10:59:42.322037 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:42 crc kubenswrapper[4846]: E0320 10:59:42.322310 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.202213 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.202477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.202636 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.202567177 +0000 UTC m=+218.568975443 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.202813 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.202862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.202996 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.202960556 +0000 UTC m=+218.569368822 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.203161 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.203347 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.203294365 +0000 UTC m=+218.569702781 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.304351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.304441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.304497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304627 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304714 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304749 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304769 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304770 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304838 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304882 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.304782 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs podName:aee42334-a152-4ac7-8778-a1bdfad832af nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.304745536 +0000 UTC m=+218.671153812 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs") pod "network-metrics-daemon-zljxc" (UID: "aee42334-a152-4ac7-8778-a1bdfad832af") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.305007 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.304976362 +0000 UTC m=+218.671384628 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.305041 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:47.305021563 +0000 UTC m=+218.671429829 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.323004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.323256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.323642 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:43 crc kubenswrapper[4846]: I0320 10:59:43.324004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.324254 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:43 crc kubenswrapper[4846]: E0320 10:59:43.324691 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:44 crc kubenswrapper[4846]: I0320 10:59:44.322426 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:44 crc kubenswrapper[4846]: E0320 10:59:44.322695 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:44 crc kubenswrapper[4846]: E0320 10:59:44.506762 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:45 crc kubenswrapper[4846]: I0320 10:59:45.321647 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:45 crc kubenswrapper[4846]: I0320 10:59:45.321683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:45 crc kubenswrapper[4846]: I0320 10:59:45.321749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:45 crc kubenswrapper[4846]: E0320 10:59:45.321819 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:45 crc kubenswrapper[4846]: E0320 10:59:45.322062 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:45 crc kubenswrapper[4846]: E0320 10:59:45.322270 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:46 crc kubenswrapper[4846]: I0320 10:59:46.321803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:46 crc kubenswrapper[4846]: E0320 10:59:46.322099 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:47 crc kubenswrapper[4846]: I0320 10:59:47.322205 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:47 crc kubenswrapper[4846]: I0320 10:59:47.322344 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:47 crc kubenswrapper[4846]: I0320 10:59:47.322205 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:47 crc kubenswrapper[4846]: E0320 10:59:47.322544 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:47 crc kubenswrapper[4846]: E0320 10:59:47.322423 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:47 crc kubenswrapper[4846]: E0320 10:59:47.322816 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:48 crc kubenswrapper[4846]: I0320 10:59:48.322216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:48 crc kubenswrapper[4846]: E0320 10:59:48.322457 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.322067 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.322206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.322091 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:49 crc kubenswrapper[4846]: E0320 10:59:49.322406 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:49 crc kubenswrapper[4846]: E0320 10:59:49.322630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:49 crc kubenswrapper[4846]: E0320 10:59:49.322830 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.355150 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.372657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.388255 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.408809 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.427521 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.460758 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.483256 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.505632 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: E0320 10:59:49.507594 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.522045 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.540145 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.554360 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.571781 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.588139 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.605175 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.619656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.631232 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.647747 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:49 crc kubenswrapper[4846]: I0320 10:59:49.662801 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:49Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:50 crc kubenswrapper[4846]: I0320 10:59:50.322046 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:50 crc kubenswrapper[4846]: E0320 10:59:50.322260 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:51 crc kubenswrapper[4846]: I0320 10:59:51.322280 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:51 crc kubenswrapper[4846]: E0320 10:59:51.322574 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:51 crc kubenswrapper[4846]: I0320 10:59:51.323041 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:51 crc kubenswrapper[4846]: I0320 10:59:51.323055 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:51 crc kubenswrapper[4846]: E0320 10:59:51.323169 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:51 crc kubenswrapper[4846]: E0320 10:59:51.323275 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:51 crc kubenswrapper[4846]: I0320 10:59:51.323466 4846 scope.go:117] "RemoveContainer" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.321736 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.321886 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.527632 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/2.log" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.530190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.531131 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.553194 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.575459 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.596374 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.613000 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.627720 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.656140 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.663695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.663740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.663752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.663767 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.663778 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:52Z","lastTransitionTime":"2026-03-20T10:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.674286 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.677271 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.682620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.682663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.682674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.682697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.682712 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:52Z","lastTransitionTime":"2026-03-20T10:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.692473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.704049 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.705795 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.710276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.710314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.710325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.710348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.710363 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:52Z","lastTransitionTime":"2026-03-20T10:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.724400 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.729063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.729126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.729142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.729170 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.729188 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:52Z","lastTransitionTime":"2026-03-20T10:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.731167 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.743233 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.743551 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.748784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.748839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.748853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.748874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.748887 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T10:59:52Z","lastTransitionTime":"2026-03-20T10:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.762298 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.765240 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a264a083-4166-477a-9480-b106826ad668\\\",\\\"systemUUID\\\":\\\"07ed3b8f-e6e3-4309-bb4a-b75c72b034f6\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: E0320 10:59:52.765402 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.774658 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.794757 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.806142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.819854 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.836640 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:52 crc kubenswrapper[4846]: I0320 10:59:52.849646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:52Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.322120 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.322120 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.322314 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:53 crc kubenswrapper[4846]: E0320 10:59:53.322505 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:53 crc kubenswrapper[4846]: E0320 10:59:53.322709 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:53 crc kubenswrapper[4846]: E0320 10:59:53.322799 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.537775 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/3.log" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.539754 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/2.log" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.544387 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" exitCode=1 Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.544476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.544550 4846 scope.go:117] "RemoveContainer" containerID="5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.545533 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 10:59:53 crc kubenswrapper[4846]: E0320 10:59:53.545763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.571436 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.587649 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.603334 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.620564 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.634634 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.653319 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.668101 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.690862 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.707680 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.722015 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.741028 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.759552 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.774121 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.784801 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.806741 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c22f4e2294b1b9e1fc6445e151477288fa89171ad6c0aee046a74dd80fd2766\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:23Z\\\",\\\"message\\\":\\\"place/certified-operators for network=default are: map[]\\\\nI0320 10:59:23.442333 7096 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 10:59:23.442325 7096 factory.go:656] Stopping watch factory\\\\nI0320 10:59:23.442400 7096 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:23.442437 7096 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0320 10:59:23.442488 7096 services_controller.go:443] Built service openshift-marketplace/certified-operators LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.214\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:50051, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0320 10:59:23.442592 7096 services_controller.go:444] Built service openshift-marketplace/certified-operators LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0320 10:59:23.442633 7096 services_controller.go:445] Built service openshift-marketplace/certified-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nF0320 10:59:23.442567 7096 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 10:59:52.642394 7414 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 10:59:52.642426 7414 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 10:59:52.642440 7414 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 10:59:52.642448 7414 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 10:59:52.642472 7414 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 10:59:52.644095 7414 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 10:59:52.644140 7414 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 10:59:52.644147 7414 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 10:59:52.644175 7414 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 10:59:52.644183 7414 factory.go:656] Stopping watch factory\\\\nI0320 10:59:52.644201 7414 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 10:59:52.644207 7414 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:52.644214 7414 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 10:59:52.644246 7414 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 10:59:52.644341 7414 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.828532 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.849775 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:53 crc kubenswrapper[4846]: I0320 10:59:53.864501 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:53Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.322781 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:54 crc kubenswrapper[4846]: E0320 10:59:54.323090 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:54 crc kubenswrapper[4846]: E0320 10:59:54.509717 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.552621 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/3.log" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.559212 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 10:59:54 crc kubenswrapper[4846]: E0320 10:59:54.559486 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.579689 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aa3c1b4-0157-4c1e-b6e1-cc4425a84cc5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8933979d3ce10dfe7d001b34bc067fb7eb5190bfac22c0641f90c0fef15f03cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39a5d3e4e479f0a7efae2b0b3fd3f013dc549f6e5976dc61b72d1c009f6e3d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.600585 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1fe6869e521e837945a2ba9931780b8e7155b4584a47bd7fdaacffd81f48b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a2181e75d2484e6d219c259e73641efdab66f1ba8bb90670e4db27b9c2da21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.617222 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d65a1dad11c778cfd880c9e662afcc640ad23b16409f79aa1bb5c610f814771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.632937 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j4h4w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9275b8c7-ce16-4c4a-a7eb-93d263b751d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://744fb1e3316590811f9fed128c0c848d6bc8b7fa7ec18dd51430347cde13ebd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msz4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j4h4w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.666659 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1a81831-258d-4fad-adc4-02dee4e744d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:52Z\\\",\\\"message\\\":\\\".go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 10:59:52.642394 7414 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 10:59:52.642426 7414 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 10:59:52.642440 7414 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 10:59:52.642448 7414 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 10:59:52.642472 7414 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 10:59:52.644095 7414 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 10:59:52.644140 7414 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 10:59:52.644147 7414 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 10:59:52.644175 7414 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 10:59:52.644183 7414 factory.go:656] Stopping watch factory\\\\nI0320 10:59:52.644201 7414 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 10:59:52.644207 7414 ovnkube.go:599] Stopped ovnkube\\\\nI0320 10:59:52.644214 7414 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 10:59:52.644246 7414 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 10:59:52.644341 7414 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:59:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv8k6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cqml5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.687281 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-cqnfg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dfe48192-f5b6-4ba5-a4a4-475459fd1815\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T10:59:39Z\\\",\\\"message\\\":\\\"2026-03-20T10:58:53+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b\\\\n2026-03-20T10:58:53+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02cb7482-6ee1-44d4-9b33-15cda822d20b to /host/opt/cni/bin/\\\\n2026-03-20T10:58:54Z [verbose] multus-daemon started\\\\n2026-03-20T10:58:54Z [verbose] Readiness Indicator file check\\\\n2026-03-20T10:59:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cplvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-cqnfg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.707698 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.726040 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.742358 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7rtn2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b2799b9e-c7c5-4816-97d6-d181a459f81f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0d17bfadfc36f8427c80dbf15b6e28ccd65385e84ed0522484be9b56eeaae31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzlmb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7rtn2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.759746 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba1949319d95dbf125b3a18aafab821ad2b32ec5a2341e691f44c0721cec71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzg6v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jds6p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.780931 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8533b04-cc98-4712-95b3-1dee5598b9b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a87ba2d4e9ba892f581403ee335d8f48ed0c23c11527ef4b2fdea1377c29a14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3435d5216f9247523e4e0c51a2f4dd42c3b373e948f7864ae6ef914ca2093274\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://986f3c55c3ce180056962beddf8ed7f06c0b71d8ca555f9e8ae4981339d143aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b62ed479895848ccefb697a217d50c1cb73c34ad1f3b8c4919433235f30a75e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b2d597f1f0cdb3dbc470c2de9dea23f237c731389159da8e385ecb5376d47de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35899811a45ec9a1ea15353c322b6ddac3145df5167ed4d9eb7884e4d2fa1007\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3914b0e89dfa36e73d33519a2f18b0a0391d65d7b860c212255b605013a59f2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xzkbs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z8m7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.798026 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zljxc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aee42334-a152-4ac7-8778-a1bdfad832af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tx2w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zljxc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.820263 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:59:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:58:19Z\\\",\\\"message\\\":\\\"g file observer\\\\nW0320 10:58:18.913596 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 10:58:18.913983 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 10:58:18.915311 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99335523/tls.crt::/tmp/serving-cert-99335523/tls.key\\\\\\\"\\\\nI0320 10:58:19.312708 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 10:58:19.314346 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 10:58:19.314364 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 10:58:19.314390 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 10:58:19.314395 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 10:58:19.318290 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 10:58:19.318395 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318653 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 10:58:19.318696 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 10:58:19.318758 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 10:58:19.318877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 10:58:19.318935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0320 10:58:19.318302 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0320 10:58:19.319967 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:58:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.839435 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e56348fc-aa30-4b35-a553-9c2a239b532d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dabca4d1b1045a86a5fa0bfa5e241169018826bf4ac92199324be2498e38dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://919f1392ec7d28b4dccb830a6a91617945a4a9930cc5625377fc3210b3cc49ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://431bbb5dd0d96486b6798cbf34348e99160b2e5b8aba4a547edcfc130691d591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://835cd31c39c2d319c618a5a057a4ba9f9d2181ea07d213338e274ecb2bbad0db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T10:57:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.862437 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.884240 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efc4111d-e483-4363-8747-3db7d11ce590\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:57:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7df2d2784a3181552dbaf476d85d8cbfa6c2103312664a77ed58333a6dce98cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ceea3a2d2beed8076fd2909e8be68027bff60fe967774b3ef5cf3601d696c8a3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T10:57:37Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 10:57:11.700504 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 10:57:11.703182 1 observer_polling.go:159] Starting file observer\\\\nI0320 10:57:11.743758 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 10:57:11.749345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 10:57:37.308147 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 10:57:37.308233 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d56e2a9d8bd433cbf25000161e9bb23a19c876828f5024704295e138019113\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8701a4cd9c1d7972fa2e756484dc8c0511c39321febbb28c0da203e863b7bd0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:57:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:57:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.909160 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b3f6f5b94b1ec95fabca40ebdf3dc36863d92febba54b4d424aa18e203726f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:54 crc kubenswrapper[4846]: I0320 10:59:54.930386 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef1b0cac-65d1-4f15-84f8-a906fa5c2967\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T10:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f698e758d6d270efa24dbc4f116f4076c4b9b4a9229bf52886b7ee6eb795cda6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d14e94909f214cf41707d4545eb95903449b646f51973d816047d41f53a8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T10:58:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cnzt6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T10:58:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tpdzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T10:59:54Z is after 2025-08-24T17:21:41Z" Mar 20 10:59:55 crc kubenswrapper[4846]: I0320 10:59:55.321552 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:55 crc kubenswrapper[4846]: I0320 10:59:55.321552 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:55 crc kubenswrapper[4846]: E0320 10:59:55.321767 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:55 crc kubenswrapper[4846]: E0320 10:59:55.321846 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:55 crc kubenswrapper[4846]: I0320 10:59:55.321582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:55 crc kubenswrapper[4846]: E0320 10:59:55.322025 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:56 crc kubenswrapper[4846]: I0320 10:59:56.322149 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:56 crc kubenswrapper[4846]: E0320 10:59:56.322353 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:56 crc kubenswrapper[4846]: I0320 10:59:56.342419 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 20 10:59:57 crc kubenswrapper[4846]: I0320 10:59:57.322672 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:57 crc kubenswrapper[4846]: I0320 10:59:57.322810 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:57 crc kubenswrapper[4846]: E0320 10:59:57.322894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:57 crc kubenswrapper[4846]: I0320 10:59:57.322951 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:57 crc kubenswrapper[4846]: E0320 10:59:57.323068 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:57 crc kubenswrapper[4846]: E0320 10:59:57.323273 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:58 crc kubenswrapper[4846]: I0320 10:59:58.322614 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 10:59:58 crc kubenswrapper[4846]: E0320 10:59:58.322887 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.321612 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.321696 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 10:59:59 crc kubenswrapper[4846]: E0320 10:59:59.321832 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.321932 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 10:59:59 crc kubenswrapper[4846]: E0320 10:59:59.322078 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 10:59:59 crc kubenswrapper[4846]: E0320 10:59:59.322187 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.361149 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podStartSLOduration=120.361106302 podStartE2EDuration="2m0.361106302s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.348155007 +0000 UTC m=+170.714563273" watchObservedRunningTime="2026-03-20 10:59:59.361106302 +0000 UTC m=+170.727514578" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.396236 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-j4h4w" podStartSLOduration=120.396200254 podStartE2EDuration="2m0.396200254s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.362086916 +0000 UTC m=+170.728495162" watchObservedRunningTime="2026-03-20 10:59:59.396200254 +0000 UTC m=+170.762608520" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.472591 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.472563566 podStartE2EDuration="3.472563566s" podCreationTimestamp="2026-03-20 10:59:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.469979784 +0000 UTC m=+170.836388090" watchObservedRunningTime="2026-03-20 10:59:59.472563566 +0000 UTC m=+170.838971822" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.473563 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-cqnfg" podStartSLOduration=120.473555191 podStartE2EDuration="2m0.473555191s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.419240963 +0000 UTC m=+170.785649199" watchObservedRunningTime="2026-03-20 10:59:59.473555191 +0000 UTC m=+170.839963437" Mar 20 10:59:59 crc kubenswrapper[4846]: E0320 10:59:59.510889 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.548049 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-z8m7m" podStartSLOduration=119.548030318 podStartE2EDuration="1m59.548030318s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.547723911 +0000 UTC m=+170.914132147" watchObservedRunningTime="2026-03-20 10:59:59.548030318 +0000 UTC m=+170.914438554" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.548316 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7rtn2" podStartSLOduration=120.548312194 podStartE2EDuration="2m0.548312194s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.526644279 +0000 UTC m=+170.893052535" watchObservedRunningTime="2026-03-20 10:59:59.548312194 +0000 UTC m=+170.914720430" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.601120 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.601098346 podStartE2EDuration="1m19.601098346s" podCreationTimestamp="2026-03-20 10:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.579506222 +0000 UTC m=+170.945914458" watchObservedRunningTime="2026-03-20 10:59:59.601098346 +0000 UTC m=+170.967506582" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.601245 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=29.601241189 podStartE2EDuration="29.601241189s" podCreationTimestamp="2026-03-20 10:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.60086848 +0000 UTC m=+170.967276706" watchObservedRunningTime="2026-03-20 10:59:59.601241189 +0000 UTC m=+170.967649425" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.633821 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=56.633788539 podStartE2EDuration="56.633788539s" podCreationTimestamp="2026-03-20 10:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.633090082 +0000 UTC m=+170.999498318" watchObservedRunningTime="2026-03-20 10:59:59.633788539 +0000 UTC m=+171.000196815" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.668103 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tpdzl" podStartSLOduration=119.668078931 podStartE2EDuration="1m59.668078931s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.667844006 +0000 UTC m=+171.034252242" watchObservedRunningTime="2026-03-20 10:59:59.668078931 +0000 UTC m=+171.034487167" Mar 20 10:59:59 crc kubenswrapper[4846]: I0320 10:59:59.713133 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=36.713107204 podStartE2EDuration="36.713107204s" podCreationTimestamp="2026-03-20 10:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 10:59:59.681799734 +0000 UTC m=+171.048207980" watchObservedRunningTime="2026-03-20 10:59:59.713107204 +0000 UTC m=+171.079515440" Mar 20 11:00:00 crc kubenswrapper[4846]: I0320 11:00:00.322627 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:00 crc kubenswrapper[4846]: E0320 11:00:00.323005 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:01 crc kubenswrapper[4846]: I0320 11:00:01.322056 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:01 crc kubenswrapper[4846]: I0320 11:00:01.322117 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:01 crc kubenswrapper[4846]: I0320 11:00:01.322223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:01 crc kubenswrapper[4846]: E0320 11:00:01.322280 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:01 crc kubenswrapper[4846]: E0320 11:00:01.322442 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:01 crc kubenswrapper[4846]: E0320 11:00:01.322637 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.321576 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:02 crc kubenswrapper[4846]: E0320 11:00:02.321790 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.822748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.822833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.822857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.822892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.822953 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T11:00:02Z","lastTransitionTime":"2026-03-20T11:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.890470 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg"] Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.891769 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.895336 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.895838 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.899462 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 11:00:02 crc kubenswrapper[4846]: I0320 11:00:02.899808 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.051169 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.051639 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.051953 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.052187 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.052436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.153933 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154244 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154650 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.154941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.156768 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.164775 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.190384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b758e0d-79e3-4cb5-b483-f0b5c460ee0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nkrrg\" (UID: \"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.212453 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.322418 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.322748 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:03 crc kubenswrapper[4846]: E0320 11:00:03.322891 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.322985 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:03 crc kubenswrapper[4846]: E0320 11:00:03.323227 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:03 crc kubenswrapper[4846]: E0320 11:00:03.323474 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.395170 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.409119 4846 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.596568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" event={"ID":"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a","Type":"ContainerStarted","Data":"a36b641b44faf4fae78d724b72e1927f05eaedb4cfe8b1abdf558b8c7de7256e"} Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.596736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" event={"ID":"4b758e0d-79e3-4cb5-b483-f0b5c460ee0a","Type":"ContainerStarted","Data":"8b3433db687d4a285c192a2614407274a33fd85e304206a86775ad0a13756091"} Mar 20 11:00:03 crc kubenswrapper[4846]: I0320 11:00:03.619664 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nkrrg" podStartSLOduration=123.619639848 podStartE2EDuration="2m3.619639848s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:03.618146082 +0000 UTC m=+174.984554358" watchObservedRunningTime="2026-03-20 11:00:03.619639848 +0000 UTC m=+174.986048114" Mar 20 11:00:04 crc kubenswrapper[4846]: I0320 11:00:04.322403 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:04 crc kubenswrapper[4846]: E0320 11:00:04.322657 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:04 crc kubenswrapper[4846]: E0320 11:00:04.512745 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:05 crc kubenswrapper[4846]: I0320 11:00:05.322527 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:05 crc kubenswrapper[4846]: I0320 11:00:05.323694 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:05 crc kubenswrapper[4846]: I0320 11:00:05.324023 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:05 crc kubenswrapper[4846]: E0320 11:00:05.324004 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:05 crc kubenswrapper[4846]: E0320 11:00:05.324185 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:05 crc kubenswrapper[4846]: E0320 11:00:05.324968 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:06 crc kubenswrapper[4846]: I0320 11:00:06.322505 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:06 crc kubenswrapper[4846]: E0320 11:00:06.322749 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:07 crc kubenswrapper[4846]: I0320 11:00:07.322303 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:07 crc kubenswrapper[4846]: I0320 11:00:07.322447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:07 crc kubenswrapper[4846]: I0320 11:00:07.322600 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:07 crc kubenswrapper[4846]: E0320 11:00:07.323162 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:07 crc kubenswrapper[4846]: E0320 11:00:07.323400 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:07 crc kubenswrapper[4846]: E0320 11:00:07.323855 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:08 crc kubenswrapper[4846]: I0320 11:00:08.321495 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:08 crc kubenswrapper[4846]: E0320 11:00:08.321746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:09 crc kubenswrapper[4846]: I0320 11:00:09.322537 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:09 crc kubenswrapper[4846]: I0320 11:00:09.322676 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:09 crc kubenswrapper[4846]: E0320 11:00:09.323630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:09 crc kubenswrapper[4846]: I0320 11:00:09.324013 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:09 crc kubenswrapper[4846]: E0320 11:00:09.324191 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:09 crc kubenswrapper[4846]: E0320 11:00:09.324877 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:09 crc kubenswrapper[4846]: I0320 11:00:09.325248 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:00:09 crc kubenswrapper[4846]: E0320 11:00:09.325445 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 11:00:09 crc kubenswrapper[4846]: E0320 11:00:09.513591 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:10 crc kubenswrapper[4846]: I0320 11:00:10.322322 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:10 crc kubenswrapper[4846]: E0320 11:00:10.322548 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:11 crc kubenswrapper[4846]: I0320 11:00:11.321836 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:11 crc kubenswrapper[4846]: E0320 11:00:11.322092 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:11 crc kubenswrapper[4846]: I0320 11:00:11.321872 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:11 crc kubenswrapper[4846]: E0320 11:00:11.322215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:11 crc kubenswrapper[4846]: I0320 11:00:11.322341 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:11 crc kubenswrapper[4846]: E0320 11:00:11.322415 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:12 crc kubenswrapper[4846]: I0320 11:00:12.321885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:12 crc kubenswrapper[4846]: E0320 11:00:12.322659 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:13 crc kubenswrapper[4846]: I0320 11:00:13.322331 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:13 crc kubenswrapper[4846]: I0320 11:00:13.322331 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:13 crc kubenswrapper[4846]: I0320 11:00:13.323175 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:13 crc kubenswrapper[4846]: E0320 11:00:13.323341 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:13 crc kubenswrapper[4846]: E0320 11:00:13.323660 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:13 crc kubenswrapper[4846]: E0320 11:00:13.323824 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:14 crc kubenswrapper[4846]: I0320 11:00:14.322399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:14 crc kubenswrapper[4846]: E0320 11:00:14.322701 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:14 crc kubenswrapper[4846]: E0320 11:00:14.516396 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:15 crc kubenswrapper[4846]: I0320 11:00:15.322478 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:15 crc kubenswrapper[4846]: I0320 11:00:15.322709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:15 crc kubenswrapper[4846]: I0320 11:00:15.322813 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:15 crc kubenswrapper[4846]: E0320 11:00:15.322928 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:15 crc kubenswrapper[4846]: E0320 11:00:15.323261 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:15 crc kubenswrapper[4846]: E0320 11:00:15.323367 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:16 crc kubenswrapper[4846]: I0320 11:00:16.322365 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:16 crc kubenswrapper[4846]: E0320 11:00:16.322568 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:17 crc kubenswrapper[4846]: I0320 11:00:17.322306 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:17 crc kubenswrapper[4846]: I0320 11:00:17.322312 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:17 crc kubenswrapper[4846]: E0320 11:00:17.322673 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:17 crc kubenswrapper[4846]: I0320 11:00:17.322726 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:17 crc kubenswrapper[4846]: E0320 11:00:17.323100 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:17 crc kubenswrapper[4846]: E0320 11:00:17.323276 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:18 crc kubenswrapper[4846]: I0320 11:00:18.321837 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:18 crc kubenswrapper[4846]: E0320 11:00:18.322333 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:19 crc kubenswrapper[4846]: I0320 11:00:19.321569 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:19 crc kubenswrapper[4846]: E0320 11:00:19.322700 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:19 crc kubenswrapper[4846]: I0320 11:00:19.322715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:19 crc kubenswrapper[4846]: I0320 11:00:19.322731 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:19 crc kubenswrapper[4846]: E0320 11:00:19.322815 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:19 crc kubenswrapper[4846]: E0320 11:00:19.323188 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:19 crc kubenswrapper[4846]: E0320 11:00:19.517110 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:20 crc kubenswrapper[4846]: I0320 11:00:20.322323 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:20 crc kubenswrapper[4846]: E0320 11:00:20.322583 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:21 crc kubenswrapper[4846]: I0320 11:00:21.321887 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:21 crc kubenswrapper[4846]: I0320 11:00:21.321992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:21 crc kubenswrapper[4846]: E0320 11:00:21.322154 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:21 crc kubenswrapper[4846]: I0320 11:00:21.322184 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:21 crc kubenswrapper[4846]: E0320 11:00:21.322458 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:21 crc kubenswrapper[4846]: E0320 11:00:21.322547 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:22 crc kubenswrapper[4846]: I0320 11:00:22.321643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:22 crc kubenswrapper[4846]: E0320 11:00:22.322192 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:22 crc kubenswrapper[4846]: I0320 11:00:22.322614 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:00:22 crc kubenswrapper[4846]: E0320 11:00:22.322802 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cqml5_openshift-ovn-kubernetes(d1a81831-258d-4fad-adc4-02dee4e744d7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" Mar 20 11:00:23 crc kubenswrapper[4846]: I0320 11:00:23.322565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:23 crc kubenswrapper[4846]: I0320 11:00:23.322689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:23 crc kubenswrapper[4846]: I0320 11:00:23.322734 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:23 crc kubenswrapper[4846]: E0320 11:00:23.322998 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:23 crc kubenswrapper[4846]: E0320 11:00:23.323247 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:23 crc kubenswrapper[4846]: E0320 11:00:23.323429 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:24 crc kubenswrapper[4846]: I0320 11:00:24.322213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:24 crc kubenswrapper[4846]: E0320 11:00:24.322702 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:24 crc kubenswrapper[4846]: E0320 11:00:24.518383 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:25 crc kubenswrapper[4846]: I0320 11:00:25.321799 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:25 crc kubenswrapper[4846]: I0320 11:00:25.321958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:25 crc kubenswrapper[4846]: E0320 11:00:25.322191 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:25 crc kubenswrapper[4846]: I0320 11:00:25.322221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:25 crc kubenswrapper[4846]: E0320 11:00:25.322741 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:25 crc kubenswrapper[4846]: E0320 11:00:25.322630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.322133 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:26 crc kubenswrapper[4846]: E0320 11:00:26.322323 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.689596 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/1.log" Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.690454 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/0.log" Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.690545 4846 generic.go:334] "Generic (PLEG): container finished" podID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" containerID="8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9" exitCode=1 Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.690653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerDied","Data":"8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9"} Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.690847 4846 scope.go:117] "RemoveContainer" containerID="98d4cd364dc8de1b71c5bbf2c7bf384dad88d3aa35468e75e9e7ed0da536f61f" Mar 20 11:00:26 crc kubenswrapper[4846]: I0320 11:00:26.691561 4846 scope.go:117] "RemoveContainer" containerID="8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9" Mar 20 11:00:26 crc kubenswrapper[4846]: E0320 11:00:26.692068 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-cqnfg_openshift-multus(dfe48192-f5b6-4ba5-a4a4-475459fd1815)\"" pod="openshift-multus/multus-cqnfg" podUID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" Mar 20 11:00:27 crc kubenswrapper[4846]: I0320 11:00:27.321674 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:27 crc kubenswrapper[4846]: I0320 11:00:27.321739 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:27 crc kubenswrapper[4846]: I0320 11:00:27.321801 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:27 crc kubenswrapper[4846]: E0320 11:00:27.321879 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:27 crc kubenswrapper[4846]: E0320 11:00:27.322000 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:27 crc kubenswrapper[4846]: E0320 11:00:27.322058 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:27 crc kubenswrapper[4846]: I0320 11:00:27.696937 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/1.log" Mar 20 11:00:28 crc kubenswrapper[4846]: I0320 11:00:28.322018 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:28 crc kubenswrapper[4846]: E0320 11:00:28.322212 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:29 crc kubenswrapper[4846]: I0320 11:00:29.321809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:29 crc kubenswrapper[4846]: I0320 11:00:29.321955 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:29 crc kubenswrapper[4846]: E0320 11:00:29.323111 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:29 crc kubenswrapper[4846]: E0320 11:00:29.323190 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:29 crc kubenswrapper[4846]: I0320 11:00:29.322047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:29 crc kubenswrapper[4846]: E0320 11:00:29.323535 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:29 crc kubenswrapper[4846]: E0320 11:00:29.518911 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:30 crc kubenswrapper[4846]: I0320 11:00:30.322116 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:30 crc kubenswrapper[4846]: E0320 11:00:30.322654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:31 crc kubenswrapper[4846]: I0320 11:00:31.321767 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:31 crc kubenswrapper[4846]: I0320 11:00:31.321870 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:31 crc kubenswrapper[4846]: I0320 11:00:31.321862 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:31 crc kubenswrapper[4846]: E0320 11:00:31.322090 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:31 crc kubenswrapper[4846]: E0320 11:00:31.322554 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:31 crc kubenswrapper[4846]: E0320 11:00:31.323260 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:32 crc kubenswrapper[4846]: I0320 11:00:32.322479 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:32 crc kubenswrapper[4846]: E0320 11:00:32.322981 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:33 crc kubenswrapper[4846]: I0320 11:00:33.321662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:33 crc kubenswrapper[4846]: I0320 11:00:33.321684 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:33 crc kubenswrapper[4846]: E0320 11:00:33.322039 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:33 crc kubenswrapper[4846]: I0320 11:00:33.321722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:33 crc kubenswrapper[4846]: E0320 11:00:33.322198 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:33 crc kubenswrapper[4846]: E0320 11:00:33.322392 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:34 crc kubenswrapper[4846]: I0320 11:00:34.321826 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:34 crc kubenswrapper[4846]: E0320 11:00:34.322093 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:34 crc kubenswrapper[4846]: E0320 11:00:34.520423 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:35 crc kubenswrapper[4846]: I0320 11:00:35.321770 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:35 crc kubenswrapper[4846]: E0320 11:00:35.321945 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:35 crc kubenswrapper[4846]: I0320 11:00:35.321887 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:35 crc kubenswrapper[4846]: E0320 11:00:35.322163 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:35 crc kubenswrapper[4846]: I0320 11:00:35.322233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:35 crc kubenswrapper[4846]: E0320 11:00:35.322287 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:36 crc kubenswrapper[4846]: I0320 11:00:36.321731 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:36 crc kubenswrapper[4846]: E0320 11:00:36.321892 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.322124 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.322125 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:37 crc kubenswrapper[4846]: E0320 11:00:37.323285 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:37 crc kubenswrapper[4846]: E0320 11:00:37.323407 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.322202 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:37 crc kubenswrapper[4846]: E0320 11:00:37.323512 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.323808 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.736573 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/3.log" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.738929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerStarted","Data":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.739628 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 11:00:37 crc kubenswrapper[4846]: I0320 11:00:37.790241 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podStartSLOduration=157.79021829 podStartE2EDuration="2m37.79021829s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:37.79020641 +0000 UTC m=+209.156614666" watchObservedRunningTime="2026-03-20 11:00:37.79021829 +0000 UTC m=+209.156626526" Mar 20 11:00:38 crc kubenswrapper[4846]: I0320 11:00:38.252674 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zljxc"] Mar 20 11:00:38 crc kubenswrapper[4846]: I0320 11:00:38.252836 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:38 crc kubenswrapper[4846]: E0320 11:00:38.252938 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:38 crc kubenswrapper[4846]: I0320 11:00:38.322410 4846 scope.go:117] "RemoveContainer" containerID="8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9" Mar 20 11:00:38 crc kubenswrapper[4846]: I0320 11:00:38.744849 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/1.log" Mar 20 11:00:38 crc kubenswrapper[4846]: I0320 11:00:38.745340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerStarted","Data":"37bdd76cbbac0fe900855b30e99d6a6d41e56b26445376aa1a01381c1438cd84"} Mar 20 11:00:39 crc kubenswrapper[4846]: I0320 11:00:39.322481 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:39 crc kubenswrapper[4846]: I0320 11:00:39.322483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:39 crc kubenswrapper[4846]: I0320 11:00:39.322618 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:39 crc kubenswrapper[4846]: E0320 11:00:39.323410 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:39 crc kubenswrapper[4846]: E0320 11:00:39.323634 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:39 crc kubenswrapper[4846]: E0320 11:00:39.323696 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:39 crc kubenswrapper[4846]: E0320 11:00:39.521660 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 11:00:40 crc kubenswrapper[4846]: I0320 11:00:40.322341 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:40 crc kubenswrapper[4846]: E0320 11:00:40.322543 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:41 crc kubenswrapper[4846]: I0320 11:00:41.323217 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:41 crc kubenswrapper[4846]: I0320 11:00:41.323329 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:41 crc kubenswrapper[4846]: E0320 11:00:41.324183 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:41 crc kubenswrapper[4846]: E0320 11:00:41.324372 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:41 crc kubenswrapper[4846]: I0320 11:00:41.323372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:41 crc kubenswrapper[4846]: E0320 11:00:41.324565 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:42 crc kubenswrapper[4846]: I0320 11:00:42.322200 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:42 crc kubenswrapper[4846]: E0320 11:00:42.322471 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:43 crc kubenswrapper[4846]: I0320 11:00:43.321767 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:43 crc kubenswrapper[4846]: I0320 11:00:43.321874 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:43 crc kubenswrapper[4846]: I0320 11:00:43.321960 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:43 crc kubenswrapper[4846]: E0320 11:00:43.322051 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 11:00:43 crc kubenswrapper[4846]: E0320 11:00:43.322246 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 11:00:43 crc kubenswrapper[4846]: E0320 11:00:43.322432 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 11:00:44 crc kubenswrapper[4846]: I0320 11:00:44.322030 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:44 crc kubenswrapper[4846]: E0320 11:00:44.322244 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zljxc" podUID="aee42334-a152-4ac7-8778-a1bdfad832af" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.321695 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.321695 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.321796 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.325833 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.325985 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.326012 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 11:00:45 crc kubenswrapper[4846]: I0320 11:00:45.326332 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 11:00:46 crc kubenswrapper[4846]: I0320 11:00:46.322170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:46 crc kubenswrapper[4846]: I0320 11:00:46.325580 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 11:00:46 crc kubenswrapper[4846]: I0320 11:00:46.325870 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.220542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.220735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:47 crc kubenswrapper[4846]: E0320 11:00:47.220831 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:02:49.220774552 +0000 UTC m=+340.587182828 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.221193 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.229607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.232575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.322509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.322590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.322643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.328110 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.328120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aee42334-a152-4ac7-8778-a1bdfad832af-metrics-certs\") pod \"network-metrics-daemon-zljxc\" (UID: \"aee42334-a152-4ac7-8778-a1bdfad832af\") " pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.340525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.448338 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.462200 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.476052 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.539925 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zljxc" Mar 20 11:00:47 crc kubenswrapper[4846]: I0320 11:00:47.807274 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zljxc"] Mar 20 11:00:47 crc kubenswrapper[4846]: W0320 11:00:47.918071 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-6b8793566e15277d529bf48a455ca5b0dddfcb19cad5f25350633d2b9db6eedd WatchSource:0}: Error finding container 6b8793566e15277d529bf48a455ca5b0dddfcb19cad5f25350633d2b9db6eedd: Status 404 returned error can't find the container with id 6b8793566e15277d529bf48a455ca5b0dddfcb19cad5f25350633d2b9db6eedd Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.790769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ad1d8af6f343a7c28e446958e3e5a4fc2d000d02902ce83284eba05279d85fb1"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.791552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d3d25bc677e5003c4101cf1475e3f7a916ae732fe701b6fd80f94762b7939e0c"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.791927 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.794196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zljxc" event={"ID":"aee42334-a152-4ac7-8778-a1bdfad832af","Type":"ContainerStarted","Data":"4cbdb9ed9e6cced4784e4e7e718c4e81163e3b1c95c537d21a8eee685742dfd5"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.794255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zljxc" event={"ID":"aee42334-a152-4ac7-8778-a1bdfad832af","Type":"ContainerStarted","Data":"d9e26bb9148126a7a1351261a6af563a36a89fc9d1cfc2e66b0951beb335d640"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.794271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zljxc" event={"ID":"aee42334-a152-4ac7-8778-a1bdfad832af","Type":"ContainerStarted","Data":"2796f01b0b77bdc24bc2e9f15dfe6269fafde1f80264ab3d92934d103e030fcf"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.796094 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e17a72338812f34edaee6c54a840223d78f313778275f557d38eb18059561ff6"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.796131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6b8793566e15277d529bf48a455ca5b0dddfcb19cad5f25350633d2b9db6eedd"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.798105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6b5a99640855abd6b5e30864aad9ba1e5fbe2a511cc1581dde130935967abf1b"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.798196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0145419b213b2fd631f4008cad06c34be84bbd8eb8ed9ab0bbf5d5900990f3e9"} Mar 20 11:00:48 crc kubenswrapper[4846]: I0320 11:00:48.865063 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zljxc" podStartSLOduration=168.865038402 podStartE2EDuration="2m48.865038402s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:48.86384753 +0000 UTC m=+220.230255776" watchObservedRunningTime="2026-03-20 11:00:48.865038402 +0000 UTC m=+220.231446658" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.848350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.906340 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4npxl"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.907207 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.908667 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sjmnj"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.910082 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.912288 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.912926 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.913582 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.914210 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:53 crc kubenswrapper[4846]: W0320 11:00:53.917938 4846 reflector.go:561] object-"openshift-authentication-operator"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Mar 20 11:00:53 crc kubenswrapper[4846]: E0320 11:00:53.918046 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 11:00:53 crc kubenswrapper[4846]: W0320 11:00:53.918405 4846 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj": failed to list *v1.Secret: secrets "authentication-operator-dockercfg-mz9bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Mar 20 11:00:53 crc kubenswrapper[4846]: E0320 11:00:53.918464 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-mz9bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"authentication-operator-dockercfg-mz9bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.921159 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.926246 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.926328 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.927700 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.928553 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.932425 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944196 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944340 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944424 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944610 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944845 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.944879 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.945131 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.945515 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.947171 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.949789 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.964474 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.964735 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.964955 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.965451 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.965584 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.965596 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.965856 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.966946 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.967703 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.968239 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.968440 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.968526 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.969150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.972217 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.972955 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.977277 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.977754 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978068 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978195 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978313 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978065 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978635 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.978965 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979001 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979162 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979180 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979244 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979303 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979291 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979342 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.979875 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.980888 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.982597 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.983367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.983966 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.984558 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.986348 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w"] Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.987667 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.988280 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.996158 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.996366 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.996677 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.996970 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.997091 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.997623 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.997980 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.998288 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.998504 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.998750 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.998948 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 11:00:53 crc kubenswrapper[4846]: I0320 11:00:53.999070 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:53.999780 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.000625 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.006939 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-encryption-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9zpl\" (UniqueName: \"kubernetes.io/projected/214e3184-d108-46d8-839b-e4cd825e8685-kube-api-access-s9zpl\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-serving-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-encryption-config\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-audit-policies\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47224603-942d-46df-9aea-caf9c98fb9e0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.007931 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008115 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-node-pullsecrets\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008170 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-client\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008202 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-etcd-client\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwxdx\" (UniqueName: \"kubernetes.io/projected/016c2ddf-70d7-43be-891c-17826c04c465-kube-api-access-lwxdx\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008261 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008296 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit-dir\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008366 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqncc\" (UniqueName: \"kubernetes.io/projected/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-kube-api-access-kqncc\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008400 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rthdr\" (UniqueName: \"kubernetes.io/projected/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-kube-api-access-rthdr\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008428 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-service-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008469 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b74tx\" (UniqueName: \"kubernetes.io/projected/47224603-942d-46df-9aea-caf9c98fb9e0-kube-api-access-b74tx\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/016c2ddf-70d7-43be-891c-17826c04c465-audit-dir\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-image-import-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008581 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47224603-942d-46df-9aea-caf9c98fb9e0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008667 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008688 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-serving-cert\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008734 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-serving-cert\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008759 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-config\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.008780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.012275 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.012698 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.012861 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.013053 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.013170 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bgm6f"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.013756 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.014291 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.015814 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.016865 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.017487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.018033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.018294 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.020656 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.020858 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.021076 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.047082 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.047410 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.047658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.048056 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.048239 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.048374 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.048522 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.048918 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.050267 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.050386 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.050775 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.050961 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-fkvwp"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.052705 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.053766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.053953 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.054226 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.055864 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.060274 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.061436 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.062136 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.063160 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sn4x4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.063992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.064197 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.066423 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.066629 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.066785 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t4dz2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.070377 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-s2rfb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.070531 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.071238 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kgqq8"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.071386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.071614 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.071878 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.071951 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.072674 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.078362 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.078573 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.078775 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.078860 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.078916 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079077 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079133 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079146 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079252 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079438 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079466 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079560 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079661 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.079793 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080035 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080194 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080303 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080409 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080756 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080792 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.080873 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.081107 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.081254 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6q4br"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.082046 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4npxl"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.082185 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.082612 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.082754 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.082921 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.092348 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.095106 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.095290 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7n8h2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.095964 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.099157 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.101493 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eeb194a-81cd-458c-9431-d34562ceba76-machine-approver-tls\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109608 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/016c2ddf-70d7-43be-891c-17826c04c465-audit-dir\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-config\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109732 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a66972d-0296-4aee-8423-8455b65edb47-proxy-tls\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9f4e46fa-3163-4d53-b27c-bc04ba14576b-proxy-tls\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-image-import-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109786 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-auth-proxy-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109805 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109823 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/287581a7-1a4c-49d3-8281-bbde95429c96-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f742b066-71a8-4af3-a3f7-dd40aa2a8504-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109910 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025e4d1b-9346-4a52-8723-278549e1f1cc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47224603-942d-46df-9aea-caf9c98fb9e0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.109678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/016c2ddf-70d7-43be-891c-17826c04c465-audit-dir\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxjq\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-kube-api-access-6rxjq\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110120 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31fde7da-f4f5-4387-9186-5102267354eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110148 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110187 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-serving-cert\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31fde7da-f4f5-4387-9186-5102267354eb-config\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110309 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-service-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110332 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54hpq\" (UniqueName: \"kubernetes.io/projected/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-kube-api-access-54hpq\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvstz\" (UniqueName: \"kubernetes.io/projected/ff4193a8-a16d-4803-b894-3c9d411677b2-kube-api-access-qvstz\") pod \"downloads-7954f5f757-fkvwp\" (UID: \"ff4193a8-a16d-4803-b894-3c9d411677b2\") " pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110386 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57wfb\" (UniqueName: \"kubernetes.io/projected/9f4e46fa-3163-4d53-b27c-bc04ba14576b-kube-api-access-57wfb\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6647m\" (UniqueName: \"kubernetes.io/projected/025e4d1b-9346-4a52-8723-278549e1f1cc-kube-api-access-6647m\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f742b066-71a8-4af3-a3f7-dd40aa2a8504-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110533 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-serving-cert\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-serving-cert\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110602 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-config\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110625 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.110769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47224603-942d-46df-9aea-caf9c98fb9e0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111045 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-image-import-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-serving-cert\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-client\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31fde7da-f4f5-4387-9186-5102267354eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111307 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.111830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.112070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-config\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.112662 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.116009 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.116166 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.117157 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.117274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025e4d1b-9346-4a52-8723-278549e1f1cc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.117596 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.117768 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118081 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118374 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-encryption-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qff8q\" (UniqueName: \"kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9zpl\" (UniqueName: \"kubernetes.io/projected/214e3184-d108-46d8-839b-e4cd825e8685-kube-api-access-s9zpl\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-serving-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-encryption-config\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118724 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a66972d-0296-4aee-8423-8455b65edb47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-audit-policies\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47224603-942d-46df-9aea-caf9c98fb9e0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287581a7-1a4c-49d3-8281-bbde95429c96-config\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118938 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51518609-11ce-4f69-b6ef-a14b399d7ea5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.118967 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287581a7-1a4c-49d3-8281-bbde95429c96-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-serving-cert\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119095 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119191 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkrt7\" (UniqueName: \"kubernetes.io/projected/3a66972d-0296-4aee-8423-8455b65edb47-kube-api-access-lkrt7\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-trusted-ca\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-images\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119304 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmggl\" (UniqueName: \"kubernetes.io/projected/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-kube-api-access-rmggl\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119355 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-node-pullsecrets\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-client\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119390 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-etcd-client\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwxdx\" (UniqueName: \"kubernetes.io/projected/016c2ddf-70d7-43be-891c-17826c04c465-kube-api-access-lwxdx\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119628 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lstf5\" (UniqueName: \"kubernetes.io/projected/3eeb194a-81cd-458c-9431-d34562ceba76-kube-api-access-lstf5\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit-dir\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqncc\" (UniqueName: \"kubernetes.io/projected/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-kube-api-access-kqncc\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rthdr\" (UniqueName: \"kubernetes.io/projected/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-kube-api-access-rthdr\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119823 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-config\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119851 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7rlv\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-kube-api-access-d7rlv\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.119998 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120050 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-service-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b74tx\" (UniqueName: \"kubernetes.io/projected/47224603-942d-46df-9aea-caf9c98fb9e0-kube-api-access-b74tx\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120095 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/279211ad-6270-4845-bf0e-75589892a27b-metrics-tls\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120119 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7748f\" (UniqueName: \"kubernetes.io/projected/7dfad023-874f-4ffc-a677-116c13612fd4-kube-api-access-7748f\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wlrg\" (UniqueName: \"kubernetes.io/projected/f742b066-71a8-4af3-a3f7-dd40aa2a8504-kube-api-access-6wlrg\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279211ad-6270-4845-bf0e-75589892a27b-trusted-ca\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.120188 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51518609-11ce-4f69-b6ef-a14b399d7ea5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.121562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-serving-cert\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.122751 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-serving-ca\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.122834 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-node-pullsecrets\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.123568 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.124213 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-service-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.124687 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.125515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/214e3184-d108-46d8-839b-e4cd825e8685-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.127021 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-audit-dir\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.130178 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-etcd-client\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.130975 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-serving-cert\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.131094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47224603-942d-46df-9aea-caf9c98fb9e0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.131391 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-encryption-config\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.133427 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.133598 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.135134 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sjmnj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.140538 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.139584 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-encryption-config\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.140000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/016c2ddf-70d7-43be-891c-17826c04c465-audit-policies\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.150525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/016c2ddf-70d7-43be-891c-17826c04c465-etcd-client\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.151434 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.151732 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.151300 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.152459 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.165182 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pc5t4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.166255 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.166776 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.166845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.166877 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.204662 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.204708 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.205066 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.205375 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.205688 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.212298 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.213599 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.214233 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.215253 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566740-46wjt"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.218017 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222250 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222886 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222923 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-client\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-serving-cert\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222961 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222980 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.222998 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31fde7da-f4f5-4387-9186-5102267354eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qff8q\" (UniqueName: \"kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223079 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025e4d1b-9346-4a52-8723-278549e1f1cc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a66972d-0296-4aee-8423-8455b65edb47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223161 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223179 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287581a7-1a4c-49d3-8281-bbde95429c96-config\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223199 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51518609-11ce-4f69-b6ef-a14b399d7ea5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223217 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-serving-cert\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223271 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287581a7-1a4c-49d3-8281-bbde95429c96-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223290 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkrt7\" (UniqueName: \"kubernetes.io/projected/3a66972d-0296-4aee-8423-8455b65edb47-kube-api-access-lkrt7\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-trusted-ca\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-images\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmggl\" (UniqueName: \"kubernetes.io/projected/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-kube-api-access-rmggl\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lstf5\" (UniqueName: \"kubernetes.io/projected/3eeb194a-81cd-458c-9431-d34562ceba76-kube-api-access-lstf5\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-config\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223460 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-images\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223499 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7rlv\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-kube-api-access-d7rlv\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223540 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223563 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jntmr\" (UniqueName: \"kubernetes.io/projected/f0ef2071-e635-4779-b52c-ced7a4d44265-kube-api-access-jntmr\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223587 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/279211ad-6270-4845-bf0e-75589892a27b-metrics-tls\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223626 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7748f\" (UniqueName: \"kubernetes.io/projected/7dfad023-874f-4ffc-a677-116c13612fd4-kube-api-access-7748f\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wlrg\" (UniqueName: \"kubernetes.io/projected/f742b066-71a8-4af3-a3f7-dd40aa2a8504-kube-api-access-6wlrg\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223699 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279211ad-6270-4845-bf0e-75589892a27b-trusted-ca\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223715 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51518609-11ce-4f69-b6ef-a14b399d7ea5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223732 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-config\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eeb194a-81cd-458c-9431-d34562ceba76-machine-approver-tls\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223833 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx5ts\" (UniqueName: \"kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a66972d-0296-4aee-8423-8455b65edb47-proxy-tls\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223867 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9f4e46fa-3163-4d53-b27c-bc04ba14576b-proxy-tls\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223884 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-auth-proxy-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223920 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/287581a7-1a4c-49d3-8281-bbde95429c96-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223954 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f742b066-71a8-4af3-a3f7-dd40aa2a8504-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.223982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smlwg\" (UniqueName: \"kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/36e90b7b-0251-46ef-8ae6-b5522e3138d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224018 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224036 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224053 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skgqs\" (UniqueName: \"kubernetes.io/projected/016f2952-29ef-43cc-aa5a-5e6a3210ce09-kube-api-access-skgqs\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxjq\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-kube-api-access-6rxjq\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224093 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31fde7da-f4f5-4387-9186-5102267354eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025e4d1b-9346-4a52-8723-278549e1f1cc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224119 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gsdtt"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224299 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a66972d-0296-4aee-8423-8455b65edb47-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025e4d1b-9346-4a52-8723-278549e1f1cc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224947 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.225478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287581a7-1a4c-49d3-8281-bbde95429c96-config\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.225615 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.225663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.225733 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.226055 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-config\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.226511 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.224129 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-serving-cert\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.227328 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51518609-11ce-4f69-b6ef-a14b399d7ea5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.227364 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-serving-cert\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.227377 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-serving-cert\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.227771 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f4e46fa-3163-4d53-b27c-bc04ba14576b-images\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.227810 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.228040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-config\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.228149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9f4e46fa-3163-4d53-b27c-bc04ba14576b-proxy-tls\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.228456 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.228597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-auth-proxy-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-trusted-ca\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f742b066-71a8-4af3-a3f7-dd40aa2a8504-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229556 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/adc3107e-afa1-4fc0-a37d-67a621f106ed-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229819 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31fde7da-f4f5-4387-9186-5102267354eb-config\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229841 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-service-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54hpq\" (UniqueName: \"kubernetes.io/projected/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-kube-api-access-54hpq\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.229884 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhwc\" (UniqueName: \"kubernetes.io/projected/36e90b7b-0251-46ef-8ae6-b5522e3138d2-kube-api-access-vmhwc\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.230004 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eeb194a-81cd-458c-9431-d34562ceba76-machine-approver-tls\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.230165 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a66972d-0296-4aee-8423-8455b65edb47-proxy-tls\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.230661 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eeb194a-81cd-458c-9431-d34562ceba76-config\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31fde7da-f4f5-4387-9186-5102267354eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231130 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-client\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvstz\" (UniqueName: \"kubernetes.io/projected/ff4193a8-a16d-4803-b894-3c9d411677b2-kube-api-access-qvstz\") pod \"downloads-7954f5f757-fkvwp\" (UID: \"ff4193a8-a16d-4803-b894-3c9d411677b2\") " pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57wfb\" (UniqueName: \"kubernetes.io/projected/9f4e46fa-3163-4d53-b27c-bc04ba14576b-kube-api-access-57wfb\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6647m\" (UniqueName: \"kubernetes.io/projected/025e4d1b-9346-4a52-8723-278549e1f1cc-kube-api-access-6647m\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31fde7da-f4f5-4387-9186-5102267354eb-config\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-config\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f742b066-71a8-4af3-a3f7-dd40aa2a8504-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025e4d1b-9346-4a52-8723-278549e1f1cc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231919 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bgm6f"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.231811 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7dfad023-874f-4ffc-a677-116c13612fd4-etcd-service-ca\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.232014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287581a7-1a4c-49d3-8281-bbde95429c96-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.232133 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.232159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7q2v\" (UniqueName: \"kubernetes.io/projected/adc3107e-afa1-4fc0-a37d-67a621f106ed-kube-api-access-b7q2v\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.232513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.233167 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.233208 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.233236 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/279211ad-6270-4845-bf0e-75589892a27b-trusted-ca\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.233848 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/51518609-11ce-4f69-b6ef-a14b399d7ea5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.233880 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sn4x4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.234172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f742b066-71a8-4af3-a3f7-dd40aa2a8504-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.235090 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t4dz2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.236161 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.237269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.238334 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.239760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/279211ad-6270-4845-bf0e-75589892a27b-metrics-tls\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.240580 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.245373 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.245795 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7dfad023-874f-4ffc-a677-116c13612fd4-serving-cert\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.246105 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.249780 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.259075 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pc5t4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.259215 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.259235 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.261664 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.265131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.266512 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.269025 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.269298 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.270522 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kgqq8"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.271619 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7zwsq"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.273173 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.273285 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.274145 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b7gtv"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.275302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.275398 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7n8h2"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.277048 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.279644 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gsdtt"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.280651 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.281936 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.283078 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.284048 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6q4br"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.285392 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.286392 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.287744 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fkvwp"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.289123 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566740-46wjt"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.290144 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7gtv"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.290632 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.291226 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7zwsq"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.292240 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fmcbh"] Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.292969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.329710 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.330123 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.333693 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334153 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334199 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-images\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334272 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jntmr\" (UniqueName: \"kubernetes.io/projected/f0ef2071-e635-4779-b52c-ced7a4d44265-kube-api-access-jntmr\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334450 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334490 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334633 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx5ts\" (UniqueName: \"kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smlwg\" (UniqueName: \"kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/36e90b7b-0251-46ef-8ae6-b5522e3138d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334800 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skgqs\" (UniqueName: \"kubernetes.io/projected/016f2952-29ef-43cc-aa5a-5e6a3210ce09-kube-api-access-skgqs\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334821 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334869 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/adc3107e-afa1-4fc0-a37d-67a621f106ed-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334957 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.334986 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhwc\" (UniqueName: \"kubernetes.io/projected/36e90b7b-0251-46ef-8ae6-b5522e3138d2-kube-api-access-vmhwc\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-config\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335083 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7q2v\" (UniqueName: \"kubernetes.io/projected/adc3107e-afa1-4fc0-a37d-67a621f106ed-kube-api-access-b7q2v\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335316 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335480 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.335571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.336210 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-images\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.336559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e90b7b-0251-46ef-8ae6-b5522e3138d2-config\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.338129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/36e90b7b-0251-46ef-8ae6-b5522e3138d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.339143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.339795 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.349762 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.374262 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.378672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.390263 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.398052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.409401 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.429651 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.438255 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.450698 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.456915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.470234 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.477976 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.490618 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.498083 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.509974 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.529945 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.549558 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.557408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.571089 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.590167 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.596874 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.610793 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.630702 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.636039 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.651627 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.670693 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.696753 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.706231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.710246 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.751605 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.770415 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.790089 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.811385 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.830693 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.850493 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.869397 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.889467 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.910211 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.930038 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.948936 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 11:00:54 crc kubenswrapper[4846]: I0320 11:00:54.970285 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.001326 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.009570 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.029930 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.049400 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.069644 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.088838 4846 request.go:700] Waited for 1.006284907s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.090861 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.110623 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.112465 4846 secret.go:188] Couldn't get secret openshift-authentication-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.112601 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert podName:214e3184-d108-46d8-839b-e4cd825e8685 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:55.612577688 +0000 UTC m=+226.978985924 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert") pod "authentication-operator-69f744f599-4npxl" (UID: "214e3184-d108-46d8-839b-e4cd825e8685") : failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.130051 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.142389 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/adc3107e-afa1-4fc0-a37d-67a621f106ed-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.150115 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.170496 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.189812 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.199000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.218255 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.227389 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.229741 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.250092 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.271312 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.289705 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.310203 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.330152 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.334883 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.335962 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert podName:f0ef2071-e635-4779-b52c-ced7a4d44265 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:55.835244026 +0000 UTC m=+227.201652272 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert") pod "catalog-operator-68c6474976-g7694" (UID: "f0ef2071-e635-4779-b52c-ced7a4d44265") : failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.336095 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.336160 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.336195 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert podName:f0ef2071-e635-4779-b52c-ced7a4d44265 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:55.836164281 +0000 UTC m=+227.202572517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert") pod "catalog-operator-68c6474976-g7694" (UID: "f0ef2071-e635-4779-b52c-ced7a4d44265") : failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: E0320 11:00:55.336252 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert podName:016f2952-29ef-43cc-aa5a-5e6a3210ce09 nodeName:}" failed. No retries permitted until 2026-03-20 11:00:55.836224623 +0000 UTC m=+227.202632899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-kwbqc" (UID: "016f2952-29ef-43cc-aa5a-5e6a3210ce09") : failed to sync secret cache: timed out waiting for the condition Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.349637 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.371391 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.390219 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.411046 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.447471 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwxdx\" (UniqueName: \"kubernetes.io/projected/016c2ddf-70d7-43be-891c-17826c04c465-kube-api-access-lwxdx\") pod \"apiserver-7bbb656c7d-kbjmt\" (UID: \"016c2ddf-70d7-43be-891c-17826c04c465\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.476712 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9zpl\" (UniqueName: \"kubernetes.io/projected/214e3184-d108-46d8-839b-e4cd825e8685-kube-api-access-s9zpl\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.487345 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b74tx\" (UniqueName: \"kubernetes.io/projected/47224603-942d-46df-9aea-caf9c98fb9e0-kube-api-access-b74tx\") pod \"openshift-apiserver-operator-796bbdcf4f-tqcnb\" (UID: \"47224603-942d-46df-9aea-caf9c98fb9e0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.518813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rthdr\" (UniqueName: \"kubernetes.io/projected/e10f5d67-8fc7-4727-9ad2-f192e8e064ef-kube-api-access-rthdr\") pod \"apiserver-76f77b778f-sjmnj\" (UID: \"e10f5d67-8fc7-4727-9ad2-f192e8e064ef\") " pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.519089 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.526531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqncc\" (UniqueName: \"kubernetes.io/projected/0a35745e-3b0d-44fa-a0d2-3f44b9f437ed-kube-api-access-kqncc\") pod \"cluster-samples-operator-665b6dd947-d2pll\" (UID: \"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.529507 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.550845 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.564485 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.573279 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.590460 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.611154 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.630410 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.651316 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.657721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.669448 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.691047 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.710579 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.729761 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.742098 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb"] Mar 20 11:00:55 crc kubenswrapper[4846]: W0320 11:00:55.748841 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47224603_942d_46df_9aea_caf9c98fb9e0.slice/crio-a586c6a9e33abad603f18a600786c35e600c79d5b47599af26a49b02c5f62fd7 WatchSource:0}: Error finding container a586c6a9e33abad603f18a600786c35e600c79d5b47599af26a49b02c5f62fd7: Status 404 returned error can't find the container with id a586c6a9e33abad603f18a600786c35e600c79d5b47599af26a49b02c5f62fd7 Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.750075 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.769596 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.770774 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.781243 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt"] Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.782039 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" Mar 20 11:00:55 crc kubenswrapper[4846]: W0320 11:00:55.788206 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod016c2ddf_70d7_43be_891c_17826c04c465.slice/crio-477d5e38027d7af240f1ada233661034856ac4b12589fe99317875bda7b2d04d WatchSource:0}: Error finding container 477d5e38027d7af240f1ada233661034856ac4b12589fe99317875bda7b2d04d: Status 404 returned error can't find the container with id 477d5e38027d7af240f1ada233661034856ac4b12589fe99317875bda7b2d04d Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.810387 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.829331 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.833846 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" event={"ID":"47224603-942d-46df-9aea-caf9c98fb9e0","Type":"ContainerStarted","Data":"a586c6a9e33abad603f18a600786c35e600c79d5b47599af26a49b02c5f62fd7"} Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.836035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" event={"ID":"016c2ddf-70d7-43be-891c-17826c04c465","Type":"ContainerStarted","Data":"477d5e38027d7af240f1ada233661034856ac4b12589fe99317875bda7b2d04d"} Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.851631 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.867078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.867186 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.867287 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.873788 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.874673 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-profile-collector-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.874766 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0ef2071-e635-4779-b52c-ced7a4d44265-srv-cert\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.877200 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/016f2952-29ef-43cc-aa5a-5e6a3210ce09-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.894496 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.928479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qff8q\" (UniqueName: \"kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q\") pod \"controller-manager-879f6c89f-2qwzz\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.945234 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wlrg\" (UniqueName: \"kubernetes.io/projected/f742b066-71a8-4af3-a3f7-dd40aa2a8504-kube-api-access-6wlrg\") pod \"kube-storage-version-migrator-operator-b67b599dd-vvnx2\" (UID: \"f742b066-71a8-4af3-a3f7-dd40aa2a8504\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.968317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.989052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7rlv\" (UniqueName: \"kubernetes.io/projected/279211ad-6270-4845-bf0e-75589892a27b-kube-api-access-d7rlv\") pod \"ingress-operator-5b745b69d9-fn6sk\" (UID: \"279211ad-6270-4845-bf0e-75589892a27b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.991980 4846 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 11:00:55 crc kubenswrapper[4846]: I0320 11:00:55.997370 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sjmnj"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.010652 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.025596 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.034008 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.040033 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.072715 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxjq\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-kube-api-access-6rxjq\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.078851 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.087514 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31fde7da-f4f5-4387-9186-5102267354eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wpg65\" (UID: \"31fde7da-f4f5-4387-9186-5102267354eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.107662 4846 request.go:700] Waited for 1.878874222s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.108325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/287581a7-1a4c-49d3-8281-bbde95429c96-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4924v\" (UID: \"287581a7-1a4c-49d3-8281-bbde95429c96\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.125182 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51518609-11ce-4f69-b6ef-a14b399d7ea5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ftt8w\" (UID: \"51518609-11ce-4f69-b6ef-a14b399d7ea5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.149864 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7748f\" (UniqueName: \"kubernetes.io/projected/7dfad023-874f-4ffc-a677-116c13612fd4-kube-api-access-7748f\") pod \"etcd-operator-b45778765-bgm6f\" (UID: \"7dfad023-874f-4ffc-a677-116c13612fd4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.176697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lstf5\" (UniqueName: \"kubernetes.io/projected/3eeb194a-81cd-458c-9431-d34562ceba76-kube-api-access-lstf5\") pod \"machine-approver-56656f9798-s4qg9\" (UID: \"3eeb194a-81cd-458c-9431-d34562ceba76\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.188585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmggl\" (UniqueName: \"kubernetes.io/projected/be4ef25a-c7c5-4158-8d16-dc6fc7a807e9-kube-api-access-rmggl\") pod \"console-operator-58897d9998-sn4x4\" (UID: \"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9\") " pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.206014 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.208984 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54hpq\" (UniqueName: \"kubernetes.io/projected/b884ddf6-9cfc-41d7-9659-dd3c86dc0337-kube-api-access-54hpq\") pod \"openshift-config-operator-7777fb866f-wpkrb\" (UID: \"b884ddf6-9cfc-41d7-9659-dd3c86dc0337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.209059 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.216922 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.229397 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkrt7\" (UniqueName: \"kubernetes.io/projected/3a66972d-0296-4aee-8423-8455b65edb47-kube-api-access-lkrt7\") pod \"machine-config-controller-84d6567774-b6szm\" (UID: \"3a66972d-0296-4aee-8423-8455b65edb47\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.229546 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.247447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvstz\" (UniqueName: \"kubernetes.io/projected/ff4193a8-a16d-4803-b894-3c9d411677b2-kube-api-access-qvstz\") pod \"downloads-7954f5f757-fkvwp\" (UID: \"ff4193a8-a16d-4803-b894-3c9d411677b2\") " pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.278749 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57wfb\" (UniqueName: \"kubernetes.io/projected/9f4e46fa-3163-4d53-b27c-bc04ba14576b-kube-api-access-57wfb\") pod \"machine-config-operator-74547568cd-wdb6b\" (UID: \"9f4e46fa-3163-4d53-b27c-bc04ba14576b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.290100 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.296384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6647m\" (UniqueName: \"kubernetes.io/projected/025e4d1b-9346-4a52-8723-278549e1f1cc-kube-api-access-6647m\") pod \"openshift-controller-manager-operator-756b6f6bc6-s6npj\" (UID: \"025e4d1b-9346-4a52-8723-278549e1f1cc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.313003 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.313443 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.323151 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.332658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.349366 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.361150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.364952 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.369996 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.373354 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.390004 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.392558 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.399611 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.401662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.410808 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.445750 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.450100 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.470583 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.484579 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.521671 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jntmr\" (UniqueName: \"kubernetes.io/projected/f0ef2071-e635-4779-b52c-ced7a4d44265-kube-api-access-jntmr\") pod \"catalog-operator-68c6474976-g7694\" (UID: \"f0ef2071-e635-4779-b52c-ced7a4d44265\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.531092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx5ts\" (UniqueName: \"kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts\") pod \"oauth-openshift-558db77b4-kgqq8\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.545300 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.555039 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smlwg\" (UniqueName: \"kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg\") pod \"marketplace-operator-79b997595-5phhj\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.567398 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.572596 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skgqs\" (UniqueName: \"kubernetes.io/projected/016f2952-29ef-43cc-aa5a-5e6a3210ce09-kube-api-access-skgqs\") pod \"package-server-manager-789f6589d5-kwbqc\" (UID: \"016f2952-29ef-43cc-aa5a-5e6a3210ce09\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.578606 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bgm6f"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.591415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhwc\" (UniqueName: \"kubernetes.io/projected/36e90b7b-0251-46ef-8ae6-b5522e3138d2-kube-api-access-vmhwc\") pod \"machine-api-operator-5694c8668f-t4dz2\" (UID: \"36e90b7b-0251-46ef-8ae6-b5522e3138d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.608682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7q2v\" (UniqueName: \"kubernetes.io/projected/adc3107e-afa1-4fc0-a37d-67a621f106ed-kube-api-access-b7q2v\") pod \"multus-admission-controller-857f4d67dd-7n8h2\" (UID: \"adc3107e-afa1-4fc0-a37d-67a621f106ed\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.630514 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.653481 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.667845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/214e3184-d108-46d8-839b-e4cd825e8685-serving-cert\") pod \"authentication-operator-69f744f599-4npxl\" (UID: \"214e3184-d108-46d8-839b-e4cd825e8685\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689169 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689306 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689346 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6x4c\" (UniqueName: \"kubernetes.io/projected/6d5e53b4-542d-4ab3-8248-dea9785709dd-kube-api-access-w6x4c\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf2hg\" (UniqueName: \"kubernetes.io/projected/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-kube-api-access-cf2hg\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689599 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689652 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689690 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ndml\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-serving-cert\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689802 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-config\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689833 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-stats-auth\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: E0320 11:00:56.689865 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.189840578 +0000 UTC m=+228.556248814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.689937 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/779260d5-a4b3-41d5-9a65-86dbb0c66845-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690038 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47wrn\" (UniqueName: \"kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-cabundle\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-default-certificate\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690115 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690175 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690200 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-metrics-certs\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690233 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r849q\" (UniqueName: \"kubernetes.io/projected/400c8113-468c-407e-a209-a5ead3d4f290-kube-api-access-r849q\") pod \"migrator-59844c95c7-j89mr\" (UID: \"400c8113-468c-407e-a209-a5ead3d4f290\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d80adab-bdc5-4c87-88af-c698df414d0f-service-ca-bundle\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690415 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690472 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpl2s\" (UniqueName: \"kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690494 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.690508 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-key\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691201 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssgnl\" (UniqueName: \"kubernetes.io/projected/1d80adab-bdc5-4c87-88af-c698df414d0f-kube-api-access-ssgnl\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691584 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-srv-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691611 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691631 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691810 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.691864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmr2n\" (UniqueName: \"kubernetes.io/projected/779260d5-a4b3-41d5-9a65-86dbb0c66845-kube-api-access-jmr2n\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.692092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxj4n\" (UniqueName: \"kubernetes.io/projected/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-kube-api-access-mxj4n\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.692119 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.692140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-metrics-tls\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.692322 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2fsv\" (UniqueName: \"kubernetes.io/projected/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-kube-api-access-c2fsv\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.713889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.726254 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.741014 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.757618 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.760327 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.785240 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.793676 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:56 crc kubenswrapper[4846]: E0320 11:00:56.793992 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.293962752 +0000 UTC m=+228.660370988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794234 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d80adab-bdc5-4c87-88af-c698df414d0f-service-ca-bundle\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794307 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-mountpoint-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48mgb\" (UniqueName: \"kubernetes.io/projected/bf2f865f-4851-452b-be24-52f92e29bef0-kube-api-access-48mgb\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794431 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f71226fc-5b20-44c9-853f-adbed19887db-tmpfs\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpl2s\" (UniqueName: \"kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794522 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-key\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794558 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794609 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/437d4e9e-f459-4b74-8182-58aa5f25df5e-config-volume\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794670 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssgnl\" (UniqueName: \"kubernetes.io/projected/1d80adab-bdc5-4c87-88af-c698df414d0f-kube-api-access-ssgnl\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794707 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-srv-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794758 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.794802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.795040 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.795123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmr2n\" (UniqueName: \"kubernetes.io/projected/779260d5-a4b3-41d5-9a65-86dbb0c66845-kube-api-access-jmr2n\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797591 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxj4n\" (UniqueName: \"kubernetes.io/projected/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-kube-api-access-mxj4n\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797656 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/437d4e9e-f459-4b74-8182-58aa5f25df5e-metrics-tls\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-metrics-tls\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797742 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2fsv\" (UniqueName: \"kubernetes.io/projected/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-kube-api-access-c2fsv\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-node-bootstrap-token\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797786 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-csi-data-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6x4c\" (UniqueName: \"kubernetes.io/projected/6d5e53b4-542d-4ab3-8248-dea9785709dd-kube-api-access-w6x4c\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.797998 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9whd\" (UniqueName: \"kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd\") pod \"auto-csr-approver-29566740-46wjt\" (UID: \"8dd31859-c683-4f1a-b167-3f20a96b6f9c\") " pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798047 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf2hg\" (UniqueName: \"kubernetes.io/projected/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-kube-api-access-cf2hg\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798499 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798561 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798582 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798636 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-apiservice-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798661 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ndml\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798679 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw2vn\" (UniqueName: \"kubernetes.io/projected/437d4e9e-f459-4b74-8182-58aa5f25df5e-kube-api-access-kw2vn\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798696 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-certs\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798716 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-serving-cert\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798736 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-config\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-stats-auth\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm897\" (UniqueName: \"kubernetes.io/projected/ad7112f5-c4c3-4a89-9b97-18e3caea5913-kube-api-access-pm897\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798804 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvk9l\" (UniqueName: \"kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798823 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-webhook-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/779260d5-a4b3-41d5-9a65-86dbb0c66845-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.798979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47wrn\" (UniqueName: \"kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799016 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-526nm\" (UniqueName: \"kubernetes.io/projected/f71226fc-5b20-44c9-853f-adbed19887db-kube-api-access-526nm\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-cabundle\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-default-certificate\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f86mp\" (UniqueName: \"kubernetes.io/projected/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-kube-api-access-f86mp\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-registration-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799255 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-cert\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-socket-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799317 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-plugins-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799365 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-metrics-certs\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.799587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r849q\" (UniqueName: \"kubernetes.io/projected/400c8113-468c-407e-a209-a5ead3d4f290-kube-api-access-r849q\") pod \"migrator-59844c95c7-j89mr\" (UID: \"400c8113-468c-407e-a209-a5ead3d4f290\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.800360 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.802028 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d80adab-bdc5-4c87-88af-c698df414d0f-service-ca-bundle\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.804279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.805196 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: E0320 11:00:56.806112 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.30609204 +0000 UTC m=+228.672500266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.806208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-config\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.806628 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.807551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.807609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-cabundle\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.808625 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.809643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.810734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.812977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.813890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-metrics-tls\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.814052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-default-certificate\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.816298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.819284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.823047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.823372 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828023 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828140 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-serving-cert\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-srv-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-stats-auth\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828046 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-signing-key\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.828060 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d80adab-bdc5-4c87-88af-c698df414d0f-metrics-certs\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.829355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.829612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/779260d5-a4b3-41d5-9a65-86dbb0c66845-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.829816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.830118 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.836393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.838358 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d5e53b4-542d-4ab3-8248-dea9785709dd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.853532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r849q\" (UniqueName: \"kubernetes.io/projected/400c8113-468c-407e-a209-a5ead3d4f290-kube-api-access-r849q\") pod \"migrator-59844c95c7-j89mr\" (UID: \"400c8113-468c-407e-a209-a5ead3d4f290\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.873436 4846 generic.go:334] "Generic (PLEG): container finished" podID="016c2ddf-70d7-43be-891c-17826c04c465" containerID="91e270401096d6cf65915de0abb1853d434a27100c6052c239af6e49ade0cb77" exitCode=0 Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.873526 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" event={"ID":"016c2ddf-70d7-43be-891c-17826c04c465","Type":"ContainerDied","Data":"91e270401096d6cf65915de0abb1853d434a27100c6052c239af6e49ade0cb77"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.873863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssgnl\" (UniqueName: \"kubernetes.io/projected/1d80adab-bdc5-4c87-88af-c698df414d0f-kube-api-access-ssgnl\") pod \"router-default-5444994796-s2rfb\" (UID: \"1d80adab-bdc5-4c87-88af-c698df414d0f\") " pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:56 crc kubenswrapper[4846]: W0320 11:00:56.879165 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31fde7da_f4f5_4387_9186_5102267354eb.slice/crio-55d4a6625a7a3d6cddce822a766a65b60dfa61a5bf4cb704eb3b416bd04c44ac WatchSource:0}: Error finding container 55d4a6625a7a3d6cddce822a766a65b60dfa61a5bf4cb704eb3b416bd04c44ac: Status 404 returned error can't find the container with id 55d4a6625a7a3d6cddce822a766a65b60dfa61a5bf4cb704eb3b416bd04c44ac Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.889312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" event={"ID":"f742b066-71a8-4af3-a3f7-dd40aa2a8504","Type":"ContainerStarted","Data":"0a1f2975b3c8e1e6fb9951a474f5e9ea9826a6e1960ccd026b94cbe68bd9261f"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.889366 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" event={"ID":"f742b066-71a8-4af3-a3f7-dd40aa2a8504","Type":"ContainerStarted","Data":"42c9ef4514508e149d5636c872d2017b71d2a5a701f06f796ed841a4dc41c1b4"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.893596 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmr2n\" (UniqueName: \"kubernetes.io/projected/779260d5-a4b3-41d5-9a65-86dbb0c66845-kube-api-access-jmr2n\") pod \"control-plane-machine-set-operator-78cbb6b69f-sv5bj\" (UID: \"779260d5-a4b3-41d5-9a65-86dbb0c66845\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:56 crc kubenswrapper[4846]: E0320 11:00:56.902337 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.40231634 +0000 UTC m=+228.768724576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902543 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902578 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/437d4e9e-f459-4b74-8182-58aa5f25df5e-metrics-tls\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902613 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-node-bootstrap-token\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-csi-data-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902680 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9whd\" (UniqueName: \"kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd\") pod \"auto-csr-approver-29566740-46wjt\" (UID: \"8dd31859-c683-4f1a-b167-3f20a96b6f9c\") " pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-apiservice-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw2vn\" (UniqueName: \"kubernetes.io/projected/437d4e9e-f459-4b74-8182-58aa5f25df5e-kube-api-access-kw2vn\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902772 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-certs\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm897\" (UniqueName: \"kubernetes.io/projected/ad7112f5-c4c3-4a89-9b97-18e3caea5913-kube-api-access-pm897\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvk9l\" (UniqueName: \"kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902833 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-webhook-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902858 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-526nm\" (UniqueName: \"kubernetes.io/projected/f71226fc-5b20-44c9-853f-adbed19887db-kube-api-access-526nm\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902884 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f86mp\" (UniqueName: \"kubernetes.io/projected/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-kube-api-access-f86mp\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-registration-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-cert\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-socket-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.902980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-plugins-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-mountpoint-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903053 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48mgb\" (UniqueName: \"kubernetes.io/projected/bf2f865f-4851-452b-be24-52f92e29bef0-kube-api-access-48mgb\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903071 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f71226fc-5b20-44c9-853f-adbed19887db-tmpfs\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/437d4e9e-f459-4b74-8182-58aa5f25df5e-config-volume\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: E0320 11:00:56.903420 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.403402759 +0000 UTC m=+228.769810995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.903884 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/437d4e9e-f459-4b74-8182-58aa5f25df5e-config-volume\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.904405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-socket-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.905275 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-plugins-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.905991 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.906816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-mountpoint-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.907253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f71226fc-5b20-44c9-853f-adbed19887db-tmpfs\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.907882 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-registration-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.908221 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad7112f5-c4c3-4a89-9b97-18e3caea5913-csi-data-dir\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.908793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.913533 4846 generic.go:334] "Generic (PLEG): container finished" podID="e10f5d67-8fc7-4727-9ad2-f192e8e064ef" containerID="a2389906c85adf2347bd8473353c3fa56b79a79f420ab47ec2f9cf3a0a56231b" exitCode=0 Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.913717 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" event={"ID":"e10f5d67-8fc7-4727-9ad2-f192e8e064ef","Type":"ContainerDied","Data":"a2389906c85adf2347bd8473353c3fa56b79a79f420ab47ec2f9cf3a0a56231b"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.913812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" event={"ID":"e10f5d67-8fc7-4727-9ad2-f192e8e064ef","Type":"ContainerStarted","Data":"36101bb5a2d9277a9f12a05ea74a6a27d020a04f029bd009e805116bb7cf1092"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.915570 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-webhook-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.915851 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-cert\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.917335 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" event={"ID":"7dfad023-874f-4ffc-a677-116c13612fd4","Type":"ContainerStarted","Data":"c73ca8129df52fa1bb30224c14795a084900963b1edf3b4f71b04b93164b862e"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.926177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/437d4e9e-f459-4b74-8182-58aa5f25df5e-metrics-tls\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.929827 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.931969 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f71226fc-5b20-44c9-853f-adbed19887db-apiservice-cert\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.933089 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fkvwp"] Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.937719 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-certs\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.947561 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.949837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bf2f865f-4851-452b-be24-52f92e29bef0-node-bootstrap-token\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.950936 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpl2s\" (UniqueName: \"kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s\") pod \"console-f9d7485db-bwgwb\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.960207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" event={"ID":"279211ad-6270-4845-bf0e-75589892a27b","Type":"ContainerStarted","Data":"800aa1d509cfc737dd424d0853d08b626ff0a82c3bba90f1948d4f108a883cd2"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.960271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" event={"ID":"279211ad-6270-4845-bf0e-75589892a27b","Type":"ContainerStarted","Data":"b5bb7a62474545d1aa2b828800ca6bd160eef94676831bbfd11a943f02b2de30"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.960286 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" event={"ID":"279211ad-6270-4845-bf0e-75589892a27b","Type":"ContainerStarted","Data":"863730e24f2ab94118f09512ccfc29400514f2262478f438ddc0f58ee93f91a8"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.969842 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6417a9fe-0232-4f96-9fcb-3212b4c3bcbf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k9lh6\" (UID: \"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.980562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxj4n\" (UniqueName: \"kubernetes.io/projected/fa67f6db-90b7-4eb4-ac3a-6e76785c92e2-kube-api-access-mxj4n\") pod \"service-ca-operator-777779d784-mdwxk\" (UID: \"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.985928 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" event={"ID":"47224603-942d-46df-9aea-caf9c98fb9e0","Type":"ContainerStarted","Data":"041cac433265992daee1bbeadb2d1a5fd56df39998532deb35ae0628baa5670f"} Mar 20 11:00:56 crc kubenswrapper[4846]: I0320 11:00:56.998617 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6x4c\" (UniqueName: \"kubernetes.io/projected/6d5e53b4-542d-4ab3-8248-dea9785709dd-kube-api-access-w6x4c\") pod \"olm-operator-6b444d44fb-g4jbg\" (UID: \"6d5e53b4-542d-4ab3-8248-dea9785709dd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.005095 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf2hg\" (UniqueName: \"kubernetes.io/projected/4f68cda7-cfdc-43b8-9a20-7057bdbd028c-kube-api-access-cf2hg\") pod \"service-ca-9c57cc56f-pc5t4\" (UID: \"4f68cda7-cfdc-43b8-9a20-7057bdbd028c\") " pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.005509 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.007472 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.507443041 +0000 UTC m=+228.873851267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.014905 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.025722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.027610 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" event={"ID":"3eeb194a-81cd-458c-9431-d34562ceba76","Type":"ContainerStarted","Data":"372af47d2c7ae42cc3c5c57afef2defbc498cf00fd2517d11a19aee8def3fe25"} Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.043742 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.045363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ndml\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.049751 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" event={"ID":"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed","Type":"ContainerStarted","Data":"6d340d411fb6fb7fbc6dc64cdb6364e1cb2b0b9dfdc6d4ccbb0ac2efc997b75a"} Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.049808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" event={"ID":"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed","Type":"ContainerStarted","Data":"f332f46c0adbb53c25bf3aa568df7224d3ddccc25f0cb812e8d33c17f2b2e22d"} Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.049824 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" event={"ID":"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed","Type":"ContainerStarted","Data":"ba51a5d2ec8587026604d2fa9a7b389ca033f5741f7509c44535dd554d1d0ff8"} Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.058601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.059095 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.064223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2fsv\" (UniqueName: \"kubernetes.io/projected/17ac0e08-c160-4a84-9f64-7ecef5dd90d4-kube-api-access-c2fsv\") pod \"dns-operator-744455d44c-6q4br\" (UID: \"17ac0e08-c160-4a84-9f64-7ecef5dd90d4\") " pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.076317 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.083554 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47wrn\" (UniqueName: \"kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn\") pod \"route-controller-manager-6576b87f9c-5lnm7\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.087273 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.103433 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.107687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.108448 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.608429461 +0000 UTC m=+228.974837697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.109669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.113845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvk9l\" (UniqueName: \"kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l\") pod \"collect-profiles-29566740-85c7z\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.128835 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.128909 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9whd\" (UniqueName: \"kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd\") pod \"auto-csr-approver-29566740-46wjt\" (UID: \"8dd31859-c683-4f1a-b167-3f20a96b6f9c\") " pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.136364 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.149551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48mgb\" (UniqueName: \"kubernetes.io/projected/bf2f865f-4851-452b-be24-52f92e29bef0-kube-api-access-48mgb\") pod \"machine-config-server-fmcbh\" (UID: \"bf2f865f-4851-452b-be24-52f92e29bef0\") " pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.176105 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.176591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-526nm\" (UniqueName: \"kubernetes.io/projected/f71226fc-5b20-44c9-853f-adbed19887db-kube-api-access-526nm\") pod \"packageserver-d55dfcdfc-nbllb\" (UID: \"f71226fc-5b20-44c9-853f-adbed19887db\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.178706 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.179191 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.195930 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.207582 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f86mp\" (UniqueName: \"kubernetes.io/projected/94edc1b0-8214-4eeb-8398-c8bac2e89fb0-kube-api-access-f86mp\") pod \"ingress-canary-b7gtv\" (UID: \"94edc1b0-8214-4eeb-8398-c8bac2e89fb0\") " pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.208195 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.208483 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.708452154 +0000 UTC m=+229.074860390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.208574 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.209056 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.70904236 +0000 UTC m=+229.075450596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.211349 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.216612 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sn4x4"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.227816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw2vn\" (UniqueName: \"kubernetes.io/projected/437d4e9e-f459-4b74-8182-58aa5f25df5e-kube-api-access-kw2vn\") pod \"dns-default-7zwsq\" (UID: \"437d4e9e-f459-4b74-8182-58aa5f25df5e\") " pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.234368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7gtv" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.235183 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm897\" (UniqueName: \"kubernetes.io/projected/ad7112f5-c4c3-4a89-9b97-18e3caea5913-kube-api-access-pm897\") pod \"csi-hostpathplugin-gsdtt\" (UID: \"ad7112f5-c4c3-4a89-9b97-18e3caea5913\") " pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.245126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fmcbh" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.276973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.279849 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7n8h2"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.310082 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.310675 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.810652087 +0000 UTC m=+229.177060323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: W0320 11:00:57.311294 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe4ef25a_c7c5_4158_8d16_dc6fc7a807e9.slice/crio-5b4e3496a8d11586dde46e3789f1a6212ecc67a0c410a14d47dc631666411d69 WatchSource:0}: Error finding container 5b4e3496a8d11586dde46e3789f1a6212ecc67a0c410a14d47dc631666411d69: Status 404 returned error can't find the container with id 5b4e3496a8d11586dde46e3789f1a6212ecc67a0c410a14d47dc631666411d69 Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.320693 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kgqq8"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.343187 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:00:57 crc kubenswrapper[4846]: W0320 11:00:57.355508 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9590f0a5_0754_484b_9f20_89e6155722d6.slice/crio-154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd WatchSource:0}: Error finding container 154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd: Status 404 returned error can't find the container with id 154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.416014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.416391 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:57.916377614 +0000 UTC m=+229.282785850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.416913 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.416962 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.416978 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4npxl"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.416987 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.445077 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t4dz2"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.498248 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.519610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.520723 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.020685913 +0000 UTC m=+229.387094179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.522976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7zwsq" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.607546 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-fn6sk" podStartSLOduration=177.6075176 podStartE2EDuration="2m57.6075176s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:57.603486521 +0000 UTC m=+228.969894757" watchObservedRunningTime="2026-03-20 11:00:57.6075176 +0000 UTC m=+228.973925856" Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.624328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.624676 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.124663033 +0000 UTC m=+229.491071269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.625320 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.731976 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.732312 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.232264732 +0000 UTC m=+229.598672968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.735481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.736061 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.236042194 +0000 UTC m=+229.602450430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.740537 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6"] Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.836931 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.837552 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.337509076 +0000 UTC m=+229.703917312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.837656 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.838095 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.338079072 +0000 UTC m=+229.704487308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: W0320 11:00:57.909541 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b910eda_ae6a_4e2e_9ce2_9a139756b29f.slice/crio-5058fafd6d57803c4c4b220f604bfd3b42e8d1ba0ed2723b35979c5f18e4f2e3 WatchSource:0}: Error finding container 5058fafd6d57803c4c4b220f604bfd3b42e8d1ba0ed2723b35979c5f18e4f2e3: Status 404 returned error can't find the container with id 5058fafd6d57803c4c4b220f604bfd3b42e8d1ba0ed2723b35979c5f18e4f2e3 Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.940278 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:57 crc kubenswrapper[4846]: E0320 11:00:57.940608 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.440567032 +0000 UTC m=+229.806975308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:57 crc kubenswrapper[4846]: I0320 11:00:57.990448 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tqcnb" podStartSLOduration=178.990429559 podStartE2EDuration="2m58.990429559s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:57.927828038 +0000 UTC m=+229.294236274" watchObservedRunningTime="2026-03-20 11:00:57.990429559 +0000 UTC m=+229.356837795" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.015336 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6q4br"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.039222 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.042873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.043343 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.543326359 +0000 UTC m=+229.909734595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: W0320 11:00:58.120732 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6417a9fe_0232_4f96_9fcb_3212b4c3bcbf.slice/crio-7d55b9c556602104707bae9d69f285fb99b99b1d1a3618a63930ec26f6864c6c WatchSource:0}: Error finding container 7d55b9c556602104707bae9d69f285fb99b99b1d1a3618a63930ec26f6864c6c: Status 404 returned error can't find the container with id 7d55b9c556602104707bae9d69f285fb99b99b1d1a3618a63930ec26f6864c6c Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.143501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" event={"ID":"016f2952-29ef-43cc-aa5a-5e6a3210ce09","Type":"ContainerStarted","Data":"e30183fe6d42c2620f034dfa8b7909f1b1942db088d5f38f055bc4855b5dee52"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.145378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.146929 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.645975784 +0000 UTC m=+230.012384020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.170322 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bwgwb" event={"ID":"7b910eda-ae6a-4e2e-9ce2-9a139756b29f","Type":"ContainerStarted","Data":"5058fafd6d57803c4c4b220f604bfd3b42e8d1ba0ed2723b35979c5f18e4f2e3"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.177284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" event={"ID":"016c2ddf-70d7-43be-891c-17826c04c465","Type":"ContainerStarted","Data":"57176fb210da70ac4da237bf895d0212f4b109ebe37dd2bdf3f0753979e54b18"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.220926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" event={"ID":"e10f5d67-8fc7-4727-9ad2-f192e8e064ef","Type":"ContainerStarted","Data":"43f9ca13f76d6be487be2ea0e2bc728872c9d5e6c5ab0bbdef3e0aa14d7f5c7a"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.247852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.248319 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.748302419 +0000 UTC m=+230.114710655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.254881 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" event={"ID":"7dfad023-874f-4ffc-a677-116c13612fd4","Type":"ContainerStarted","Data":"e32f9443f78f83fa5fd33e5a7b36d641ba5a1e5c53d092560c6c0942a135c3b3"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.270163 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" podStartSLOduration=179.270137209 podStartE2EDuration="2m59.270137209s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:58.269368099 +0000 UTC m=+229.635776325" watchObservedRunningTime="2026-03-20 11:00:58.270137209 +0000 UTC m=+229.636545465" Mar 20 11:00:58 crc kubenswrapper[4846]: W0320 11:00:58.271948 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod400c8113_468c_407e_a209_a5ead3d4f290.slice/crio-1e1dabad4949753eca592bc120a8ed66c8359f726c84b83e0ec1e0bab6036374 WatchSource:0}: Error finding container 1e1dabad4949753eca592bc120a8ed66c8359f726c84b83e0ec1e0bab6036374: Status 404 returned error can't find the container with id 1e1dabad4949753eca592bc120a8ed66c8359f726c84b83e0ec1e0bab6036374 Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.274055 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" event={"ID":"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9","Type":"ContainerStarted","Data":"5b4e3496a8d11586dde46e3789f1a6212ecc67a0c410a14d47dc631666411d69"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.299997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" event={"ID":"214e3184-d108-46d8-839b-e4cd825e8685","Type":"ContainerStarted","Data":"8be7b7d7191c91987062e39a4aba0cd174ab7ec00c4892eeb1075b09fee292fe"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.351840 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.352266 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.852229808 +0000 UTC m=+230.218638044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.353224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.355991 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.85597376 +0000 UTC m=+230.222381996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.361367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" event={"ID":"3eeb194a-81cd-458c-9431-d34562ceba76","Type":"ContainerStarted","Data":"4fd156c8660f423f5911bd7369ff18b3b6b427d0c6a1aa8949cdbd0ed4908c67"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.379129 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vvnx2" podStartSLOduration=178.379092994 podStartE2EDuration="2m58.379092994s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:58.370864832 +0000 UTC m=+229.737273068" watchObservedRunningTime="2026-03-20 11:00:58.379092994 +0000 UTC m=+229.745501230" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.403241 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" event={"ID":"51518609-11ce-4f69-b6ef-a14b399d7ea5","Type":"ContainerStarted","Data":"b8c6a277b7284a7c797e02378fb8d5ed70bf56145ae8231630439cb6b3040885"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.403321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" event={"ID":"51518609-11ce-4f69-b6ef-a14b399d7ea5","Type":"ContainerStarted","Data":"6ff887afc18052895d5d22de03811c7ae8040fc4641af708e24f620f25794fb7"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.427993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.457254 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.458018 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:58.957994436 +0000 UTC m=+230.324402672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.469198 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7zwsq"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.485674 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pc5t4"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.515087 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.518375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" event={"ID":"90413730-bfef-4d28-9c45-7aa031e5af17","Type":"ContainerStarted","Data":"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.518476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" event={"ID":"90413730-bfef-4d28-9c45-7aa031e5af17","Type":"ContainerStarted","Data":"b704a5db648bc3de6c872bb5a22cfc70130bff4863813e921cef354ac7650ecd"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.519808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.528860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" event={"ID":"36e90b7b-0251-46ef-8ae6-b5522e3138d2","Type":"ContainerStarted","Data":"2016d6cc0bafb24cf1e02d5c9c1a6cd43a0b4b93b5e0f981315a4e8d15b270ec"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.545329 4846 generic.go:334] "Generic (PLEG): container finished" podID="b884ddf6-9cfc-41d7-9659-dd3c86dc0337" containerID="c179156931d13ebb7c9731edface9b2a5d10da8f71db4193308e9d734f6411f6" exitCode=0 Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.545449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" event={"ID":"b884ddf6-9cfc-41d7-9659-dd3c86dc0337","Type":"ContainerDied","Data":"c179156931d13ebb7c9731edface9b2a5d10da8f71db4193308e9d734f6411f6"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.545496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" event={"ID":"b884ddf6-9cfc-41d7-9659-dd3c86dc0337","Type":"ContainerStarted","Data":"860fc89ed8f44079dd0f0a2af21b7ea46722fdb334aac053d68f1e5b623b3aed"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.549274 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.560123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.561558 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.061543465 +0000 UTC m=+230.427951701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.581509 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fmcbh" event={"ID":"bf2f865f-4851-452b-be24-52f92e29bef0","Type":"ContainerStarted","Data":"5885bd0252ff5acb865b01bf6551cf05b632dcbdcedb0dde07590de1f0f878d0"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.599354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.600100 4846 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qwzz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.600170 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.622160 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" event={"ID":"025e4d1b-9346-4a52-8723-278549e1f1cc","Type":"ContainerStarted","Data":"4f8096fbe8b6c2a81b3035b7dfda1f3a96da6d4a5982891c48badc0c4be745d5"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.652217 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" event={"ID":"287581a7-1a4c-49d3-8281-bbde95429c96","Type":"ContainerStarted","Data":"c5a94dfa6da9d2d2b9987262c83955601265aed58c79f6abc4cfa0e771c7ba3f"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.670111 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.670666 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.170646074 +0000 UTC m=+230.537054300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.671478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" event={"ID":"f0ef2071-e635-4779-b52c-ced7a4d44265","Type":"ContainerStarted","Data":"513608ef35b971d4aa73ea69bb619a7bdcccbd26597af8302c9ddb4da0753167"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.725055 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.725482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" event={"ID":"3a66972d-0296-4aee-8423-8455b65edb47","Type":"ContainerStarted","Data":"07fc82015204967d9caa3f968f1725b324e4767aea4ce3023a4d7f4659f34090"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.740397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s2rfb" event={"ID":"1d80adab-bdc5-4c87-88af-c698df414d0f","Type":"ContainerStarted","Data":"5cfd5e408306c89985c81e119023b0a5ae66b08df07810861cdae0dcc35010cd"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.746973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" event={"ID":"4acf8a6b-4d95-435f-a697-49a803be1454","Type":"ContainerStarted","Data":"3487b6391fa123c3205072262dd4a0c7aae12a82cf3715d7d6d489d862550368"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.760112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" event={"ID":"31fde7da-f4f5-4387-9186-5102267354eb","Type":"ContainerStarted","Data":"23e1bf9fb879ecd26cff191b50922233c8237772566f78654499f6fbe35a734b"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.760162 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" event={"ID":"31fde7da-f4f5-4387-9186-5102267354eb","Type":"ContainerStarted","Data":"55d4a6625a7a3d6cddce822a766a65b60dfa61a5bf4cb704eb3b416bd04c44ac"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.766565 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerStarted","Data":"154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.767653 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.773753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.774420 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.274402188 +0000 UTC m=+230.640810424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.801366 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5phhj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.801449 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.818736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" event={"ID":"9f4e46fa-3163-4d53-b27c-bc04ba14576b","Type":"ContainerStarted","Data":"22a1e8b29db18f264f66163723b66803f054df47b7e7e96bea60a0524a4cca0e"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.842575 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566740-46wjt"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.844878 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fkvwp" event={"ID":"ff4193a8-a16d-4803-b894-3c9d411677b2","Type":"ContainerStarted","Data":"aa167f3c156049db5f0ab7d7893994fe2e0d4c37ec840b4526f50083d74a3004"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.844939 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fkvwp" event={"ID":"ff4193a8-a16d-4803-b894-3c9d411677b2","Type":"ContainerStarted","Data":"565793f59218d9a6559e4f21c7673e3d7e0abc2190d4b305f90f163bf7f56ef9"} Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.845975 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.867383 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-fkvwp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.867467 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fkvwp" podUID="ff4193a8-a16d-4803-b894-3c9d411677b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.876605 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.877917 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.377884715 +0000 UTC m=+230.744292941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:58 crc kubenswrapper[4846]: W0320 11:00:58.893675 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod437d4e9e_f459_4b74_8182_58aa5f25df5e.slice/crio-3eb60764bdfbf215f8e5f19b959dcf15381bdd07640e20ef0006c993adc278fe WatchSource:0}: Error finding container 3eb60764bdfbf215f8e5f19b959dcf15381bdd07640e20ef0006c993adc278fe: Status 404 returned error can't find the container with id 3eb60764bdfbf215f8e5f19b959dcf15381bdd07640e20ef0006c993adc278fe Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.905478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" event={"ID":"adc3107e-afa1-4fc0-a37d-67a621f106ed","Type":"ContainerStarted","Data":"7fda52c4344b91ab3276e0af14c33fc21fdd978f707b9dec5ed00a4b4397b7db"} Mar 20 11:00:58 crc kubenswrapper[4846]: W0320 11:00:58.905831 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d5e53b4_542d_4ab3_8248_dea9785709dd.slice/crio-ad36daad479a6472cfc02056662835882a013588dd9feae3977b42162173047e WatchSource:0}: Error finding container ad36daad479a6472cfc02056662835882a013588dd9feae3977b42162173047e: Status 404 returned error can't find the container with id ad36daad479a6472cfc02056662835882a013588dd9feae3977b42162173047e Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.962370 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-bgm6f" podStartSLOduration=178.962344578 podStartE2EDuration="2m58.962344578s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:58.863629 +0000 UTC m=+230.230037236" watchObservedRunningTime="2026-03-20 11:00:58.962344578 +0000 UTC m=+230.328752814" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.984634 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7gtv"] Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.985075 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" podStartSLOduration=178.985058882 podStartE2EDuration="2m58.985058882s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:58.979672546 +0000 UTC m=+230.346080782" watchObservedRunningTime="2026-03-20 11:00:58.985058882 +0000 UTC m=+230.351467118" Mar 20 11:00:58 crc kubenswrapper[4846]: I0320 11:00:58.985743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:58 crc kubenswrapper[4846]: E0320 11:00:58.986280 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.486265135 +0000 UTC m=+230.852673371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.015775 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ftt8w" podStartSLOduration=179.015750452 podStartE2EDuration="2m59.015750452s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.015302619 +0000 UTC m=+230.381710865" watchObservedRunningTime="2026-03-20 11:00:59.015750452 +0000 UTC m=+230.382158688" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.060373 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wpg65" podStartSLOduration=179.060352337 podStartE2EDuration="2m59.060352337s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.054845288 +0000 UTC m=+230.421253524" watchObservedRunningTime="2026-03-20 11:00:59.060352337 +0000 UTC m=+230.426760573" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.094512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.094983 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.594957032 +0000 UTC m=+230.961365268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.113818 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" podStartSLOduration=179.113795352 podStartE2EDuration="2m59.113795352s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.09264173 +0000 UTC m=+230.459049966" watchObservedRunningTime="2026-03-20 11:00:59.113795352 +0000 UTC m=+230.480203578" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.117077 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk"] Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.121988 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" podStartSLOduration=179.121964103 podStartE2EDuration="2m59.121964103s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.120304777 +0000 UTC m=+230.486713023" watchObservedRunningTime="2026-03-20 11:00:59.121964103 +0000 UTC m=+230.488372339" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.144270 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:00:59 crc kubenswrapper[4846]: W0320 11:00:59.147619 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa67f6db_90b7_4eb4_ac3a_6e76785c92e2.slice/crio-4fff9132473ccc206c8e33e74b770141f282fa6eb43e160313b86bbfc22364df WatchSource:0}: Error finding container 4fff9132473ccc206c8e33e74b770141f282fa6eb43e160313b86bbfc22364df: Status 404 returned error can't find the container with id 4fff9132473ccc206c8e33e74b770141f282fa6eb43e160313b86bbfc22364df Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.161692 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-fkvwp" podStartSLOduration=179.161631394 podStartE2EDuration="2m59.161631394s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.159501116 +0000 UTC m=+230.525909352" watchObservedRunningTime="2026-03-20 11:00:59.161631394 +0000 UTC m=+230.528039620" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.206474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.206803 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.706790825 +0000 UTC m=+231.073199061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.211510 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gsdtt"] Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.308151 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.308355 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.808322549 +0000 UTC m=+231.174730785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.310732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.311146 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.811132435 +0000 UTC m=+231.177540671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.422248 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.423086 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:00:59.923063441 +0000 UTC m=+231.289471687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.524813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.525433 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.025416477 +0000 UTC m=+231.391824713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.637595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.637915 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.137880096 +0000 UTC m=+231.504288332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.740575 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.744227 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.24419978 +0000 UTC m=+231.610608016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.843774 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.844234 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.344218083 +0000 UTC m=+231.710626319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.913044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" event={"ID":"ac8d3b56-111a-440c-a3c3-9ff9708af766","Type":"ContainerStarted","Data":"a99b79b89fc3031ff4e601c03e7e95714d8ec971a7d7f318a53a4467048b60ad"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.926609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerStarted","Data":"7323dedc584f0a1300da63771d0bd4af4e018a00c6e7ae60fbb8d1b55efa548d"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.927351 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5phhj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.927391 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.931349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7zwsq" event={"ID":"437d4e9e-f459-4b74-8182-58aa5f25df5e","Type":"ContainerStarted","Data":"3eb60764bdfbf215f8e5f19b959dcf15381bdd07640e20ef0006c993adc278fe"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.946018 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:00:59 crc kubenswrapper[4846]: E0320 11:00:59.946628 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.44659679 +0000 UTC m=+231.813005026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.947993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" event={"ID":"3eeb194a-81cd-458c-9431-d34562ceba76","Type":"ContainerStarted","Data":"622ce9c5117d52f03d80cd9517fc792ab257c804ad9cbaf71df99b023decab17"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.957148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" event={"ID":"4f68cda7-cfdc-43b8-9a20-7057bdbd028c","Type":"ContainerStarted","Data":"699314f23a2ed62eb7da2a4d52c21cde6ce2ea081b1c5a083502bf59dc9e9d61"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.966658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fmcbh" event={"ID":"bf2f865f-4851-452b-be24-52f92e29bef0","Type":"ContainerStarted","Data":"4d56df5423b5bacf737f0222b59a92507473babf548d56c516266ad25d065d04"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.971647 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-s4qg9" podStartSLOduration=180.971612636 podStartE2EDuration="3m0.971612636s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:00:59.965008628 +0000 UTC m=+231.331416864" watchObservedRunningTime="2026-03-20 11:00:59.971612636 +0000 UTC m=+231.338020882" Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.980036 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" event={"ID":"17ac0e08-c160-4a84-9f64-7ecef5dd90d4","Type":"ContainerStarted","Data":"3f29d2447a0cbea6034317e37f5754435bf43889b6e07af3a7e62fa95b83fb13"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.989202 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" event={"ID":"53ec1a91-8224-4f4e-a4ed-d0ee907b0354","Type":"ContainerStarted","Data":"05fae997e711156d655a2f7612c92b2e5c2bb6414cbf70f2ce52d3cdfe75410c"} Mar 20 11:00:59 crc kubenswrapper[4846]: I0320 11:00:59.998690 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" event={"ID":"f71226fc-5b20-44c9-853f-adbed19887db","Type":"ContainerStarted","Data":"bd1fd52017c13245c7d6e841909f58684bed4e73ac95af5f91bdb583de2dca2a"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.003720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" event={"ID":"287581a7-1a4c-49d3-8281-bbde95429c96","Type":"ContainerStarted","Data":"7a69f42dff63150b03eedd560f5df39229a66007d03da08725fb7583dde15af6"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.007979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" event={"ID":"025e4d1b-9346-4a52-8723-278549e1f1cc","Type":"ContainerStarted","Data":"739cd25efff93a9179456eb736d0429bd56b831cb64f7db41296e3898a3ab2c9"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.016404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" event={"ID":"adc3107e-afa1-4fc0-a37d-67a621f106ed","Type":"ContainerStarted","Data":"0bbdeaad3abd42846e81ed33caf6bc4b426af8f28001714c3e7c9af7c1e8540b"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.020536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" event={"ID":"f0ef2071-e635-4779-b52c-ced7a4d44265","Type":"ContainerStarted","Data":"715e20c094eb553183ad04e54140c3165a0eac151344b1a640e276712246479f"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.021659 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.023176 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7gtv" event={"ID":"94edc1b0-8214-4eeb-8398-c8bac2e89fb0","Type":"ContainerStarted","Data":"a5fc66024c3113b6967afab2fc7aaeb3deb49419933eb2ca9204b23e2620d8e4"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.025702 4846 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-g7694 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.025751 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" podUID="f0ef2071-e635-4779-b52c-ced7a4d44265" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.030464 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" event={"ID":"400c8113-468c-407e-a209-a5ead3d4f290","Type":"ContainerStarted","Data":"9cea609d365acdc03939d8fcb56f6aca5280c5f51e43e045f5ae31cd3c676819"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.030531 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" event={"ID":"400c8113-468c-407e-a209-a5ead3d4f290","Type":"ContainerStarted","Data":"1e1dabad4949753eca592bc120a8ed66c8359f726c84b83e0ec1e0bab6036374"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.038743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" event={"ID":"be4ef25a-c7c5-4158-8d16-dc6fc7a807e9","Type":"ContainerStarted","Data":"08014c58f29a2e5d2fa25c14dd00eabe398ccbce309d0993e06918820f12bad3"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.039162 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.040980 4846 patch_prober.go:28] interesting pod/console-operator-58897d9998-sn4x4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.041019 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" podUID="be4ef25a-c7c5-4158-8d16-dc6fc7a807e9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.042796 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4924v" podStartSLOduration=180.042783349 podStartE2EDuration="3m0.042783349s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.037336072 +0000 UTC m=+231.403744318" watchObservedRunningTime="2026-03-20 11:01:00.042783349 +0000 UTC m=+231.409191585" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.047354 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.049510 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.549483491 +0000 UTC m=+231.915891727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.076329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" event={"ID":"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf","Type":"ContainerStarted","Data":"7d55b9c556602104707bae9d69f285fb99b99b1d1a3618a63930ec26f6864c6c"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.086922 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" podStartSLOduration=180.086880092 podStartE2EDuration="3m0.086880092s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.076355467 +0000 UTC m=+231.442763703" watchObservedRunningTime="2026-03-20 11:01:00.086880092 +0000 UTC m=+231.453288328" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.093916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566740-46wjt" event={"ID":"8dd31859-c683-4f1a-b167-3f20a96b6f9c","Type":"ContainerStarted","Data":"79cbc8a2bac7be21f4034f30a0c3e022fef5875c9ff2c518852bca604f266f33"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.114532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" event={"ID":"ad7112f5-c4c3-4a89-9b97-18e3caea5913","Type":"ContainerStarted","Data":"41c3c698f92e0b065e3eb58c05ebe4cbd07808eb76f4542be5d8b3c38d0ef868"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.130414 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s6npj" podStartSLOduration=180.130391888 podStartE2EDuration="3m0.130391888s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.122364311 +0000 UTC m=+231.488772547" watchObservedRunningTime="2026-03-20 11:01:00.130391888 +0000 UTC m=+231.496800124" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.137029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" event={"ID":"016f2952-29ef-43cc-aa5a-5e6a3210ce09","Type":"ContainerStarted","Data":"3f6228b091842d63be4e45f4d29556c7e4217c3e70ad1547ba1b129ba836c660"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.149798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.150409 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.650386588 +0000 UTC m=+232.016794874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.164964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" event={"ID":"779260d5-a4b3-41d5-9a65-86dbb0c66845","Type":"ContainerStarted","Data":"16be926056ed4d5dad7f9949e79374b059f561a379101fdf7e3bf5f01e806940"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.200914 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" podStartSLOduration=180.200861632 podStartE2EDuration="3m0.200861632s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.193843742 +0000 UTC m=+231.560251998" watchObservedRunningTime="2026-03-20 11:01:00.200861632 +0000 UTC m=+231.567269868" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.253870 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.254603 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.754578035 +0000 UTC m=+232.120986271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.255835 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.256658 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.75664444 +0000 UTC m=+232.123052676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.263498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-s2rfb" event={"ID":"1d80adab-bdc5-4c87-88af-c698df414d0f","Type":"ContainerStarted","Data":"3b71a7c574ebef9bf386d5d479c57611175b64b02808c0cda51f3e90630b4197"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.275386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" event={"ID":"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2","Type":"ContainerStarted","Data":"4fff9132473ccc206c8e33e74b770141f282fa6eb43e160313b86bbfc22364df"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.356647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" event={"ID":"9f4e46fa-3163-4d53-b27c-bc04ba14576b","Type":"ContainerStarted","Data":"a62c738ef65aacba4f414ea9431464f2222aa79a4746561210b011e7c8a5aba8"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.357392 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.362723 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.862697857 +0000 UTC m=+232.229106313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.391405 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-s2rfb" podStartSLOduration=180.391383021 podStartE2EDuration="3m0.391383021s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.387393863 +0000 UTC m=+231.753802109" watchObservedRunningTime="2026-03-20 11:01:00.391383021 +0000 UTC m=+231.757791267" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.402267 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" event={"ID":"3a66972d-0296-4aee-8423-8455b65edb47","Type":"ContainerStarted","Data":"19fbfc07b3086658f95f082a8f0723fe0a1ecc0e49ba515ada4132b24e302ef8"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.412396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" event={"ID":"214e3184-d108-46d8-839b-e4cd825e8685","Type":"ContainerStarted","Data":"d5438c42cc2856749869f053753ab91fad1b5429963f0b5fd71e6687822d9147"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.440269 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4npxl" podStartSLOduration=181.440248982 podStartE2EDuration="3m1.440248982s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:00.440012766 +0000 UTC m=+231.806421002" watchObservedRunningTime="2026-03-20 11:01:00.440248982 +0000 UTC m=+231.806657228" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.460331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.460722 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:00.960706125 +0000 UTC m=+232.327114351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.477820 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" event={"ID":"6d5e53b4-542d-4ab3-8248-dea9785709dd","Type":"ContainerStarted","Data":"ad36daad479a6472cfc02056662835882a013588dd9feae3977b42162173047e"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.499578 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" event={"ID":"36e90b7b-0251-46ef-8ae6-b5522e3138d2","Type":"ContainerStarted","Data":"ffaa8e5bc6cb0ad8ef9b6e2cbd748095c90dba574bdae3dabcf065eb72ad457b"} Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.511085 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-fkvwp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.511169 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fkvwp" podUID="ff4193a8-a16d-4803-b894-3c9d411677b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.520973 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.569831 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.570022 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.570106 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.570345 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.070324798 +0000 UTC m=+232.436733084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.570850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.579069 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.079050504 +0000 UTC m=+232.445458740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.607600 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.672504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.673037 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.173019653 +0000 UTC m=+232.539427889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.776843 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.777405 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.277386024 +0000 UTC m=+232.643794260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.878650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.878989 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.378887987 +0000 UTC m=+232.745296223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.879405 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.879850 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.379831183 +0000 UTC m=+232.746239409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:00 crc kubenswrapper[4846]: I0320 11:01:00.981820 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:00 crc kubenswrapper[4846]: E0320 11:01:00.982134 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.482068916 +0000 UTC m=+232.848477152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.030735 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.042175 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:01 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:01 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:01 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.042254 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.096955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.097605 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.597586479 +0000 UTC m=+232.963994715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.200687 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.201045 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.701005704 +0000 UTC m=+233.067413940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.201393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.202061 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.702041122 +0000 UTC m=+233.068449358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.303257 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.303513 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.803481474 +0000 UTC m=+233.169889710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.303631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.304193 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.804175582 +0000 UTC m=+233.170583818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.405184 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.405920 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:01.905883701 +0000 UTC m=+233.272291937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.513455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.513860 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.013845699 +0000 UTC m=+233.380253945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.537790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" event={"ID":"400c8113-468c-407e-a209-a5ead3d4f290","Type":"ContainerStarted","Data":"7b3535c44f5961b91e3eb5f96613ed5c0c5574e6760a807f35fc296070e41593"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.543788 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" event={"ID":"6417a9fe-0232-4f96-9fcb-3212b4c3bcbf","Type":"ContainerStarted","Data":"6703bceae8be8bdb2c97f9c50e6869a6edb6b7eb201cbd133d868390f22a01f3"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.567373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" event={"ID":"4acf8a6b-4d95-435f-a697-49a803be1454","Type":"ContainerStarted","Data":"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.568585 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.579714 4846 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kgqq8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" start-of-body= Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.579766 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.582074 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j89mr" podStartSLOduration=181.582057083 podStartE2EDuration="3m1.582057083s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.579407331 +0000 UTC m=+232.945815567" watchObservedRunningTime="2026-03-20 11:01:01.582057083 +0000 UTC m=+232.948465319" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.603388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7zwsq" event={"ID":"437d4e9e-f459-4b74-8182-58aa5f25df5e","Type":"ContainerStarted","Data":"253075b6b522b3eb7e07337af2d61353f96f49680f7e5fd22f86d67322d31043"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.605991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7gtv" event={"ID":"94edc1b0-8214-4eeb-8398-c8bac2e89fb0","Type":"ContainerStarted","Data":"34b01986189a6f0806d03c8bd7750bc9022857e752d0760a26f22aa88177bb44"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.614974 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.616100 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.116075312 +0000 UTC m=+233.482483538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.616655 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.617202 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.117191003 +0000 UTC m=+233.483599239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.625657 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k9lh6" podStartSLOduration=181.625635461 podStartE2EDuration="3m1.625635461s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.624923721 +0000 UTC m=+232.991331957" watchObservedRunningTime="2026-03-20 11:01:01.625635461 +0000 UTC m=+232.992043697" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.636946 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" event={"ID":"36e90b7b-0251-46ef-8ae6-b5522e3138d2","Type":"ContainerStarted","Data":"dcf8daabc67f6fa381187c698300f8725940c77ecf10074591d6602650f00b9d"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.672130 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" event={"ID":"53ec1a91-8224-4f4e-a4ed-d0ee907b0354","Type":"ContainerStarted","Data":"5203b292d231b364464cd68a9277c86c387901b379f9a0fde9581e81195e5c86"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.685623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" event={"ID":"016f2952-29ef-43cc-aa5a-5e6a3210ce09","Type":"ContainerStarted","Data":"fc6cf4116a3033d2ab18b29c28a697670f012b543149ce60919cf482b72421b3"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.686387 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.706283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" event={"ID":"4f68cda7-cfdc-43b8-9a20-7057bdbd028c","Type":"ContainerStarted","Data":"9afcfdcc9fa57c64254aae2149f5b02aef6fedc9644c3a770c9232529824c7c2"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.719795 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.720129 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.220107094 +0000 UTC m=+233.586515340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.720402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.722000 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.221987865 +0000 UTC m=+233.588396301 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.745582 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" event={"ID":"e10f5d67-8fc7-4727-9ad2-f192e8e064ef","Type":"ContainerStarted","Data":"fe3488fdc47bd24414339a768ac378b706a3dbe6e96b3a30221cbd458d5a4d16"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.786341 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" event={"ID":"adc3107e-afa1-4fc0-a37d-67a621f106ed","Type":"ContainerStarted","Data":"70c5c43b42307175530ef7a97a1346d5d1655762abb5b9e24c3ccaf5a1298bca"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.789733 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" event={"ID":"17ac0e08-c160-4a84-9f64-7ecef5dd90d4","Type":"ContainerStarted","Data":"7058651cb8d83ec1db96701420a6c1bbfd1084f217242f25fd9de65234e2c988"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.800614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" event={"ID":"b884ddf6-9cfc-41d7-9659-dd3c86dc0337","Type":"ContainerStarted","Data":"7c6917a9bf86f4274b025b7cfe79c5e63b1e8d2ea2713f17ca7cf9836254795f"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.801488 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.822246 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.823957 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.32393417 +0000 UTC m=+233.690342416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.832571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" event={"ID":"ac8d3b56-111a-440c-a3c3-9ff9708af766","Type":"ContainerStarted","Data":"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.833935 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.839310 4846 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5lnm7 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.839374 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.840230 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" podStartSLOduration=182.84019358 podStartE2EDuration="3m2.84019358s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.746534778 +0000 UTC m=+233.112943034" watchObservedRunningTime="2026-03-20 11:01:01.84019358 +0000 UTC m=+233.206601826" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.840747 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" podStartSLOduration=61.840739124 podStartE2EDuration="1m1.840739124s" podCreationTimestamp="2026-03-20 11:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.83428374 +0000 UTC m=+233.200691986" watchObservedRunningTime="2026-03-20 11:01:01.840739124 +0000 UTC m=+233.207147360" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.855810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" event={"ID":"779260d5-a4b3-41d5-9a65-86dbb0c66845","Type":"ContainerStarted","Data":"1a97de6968ea508172ff77fe4ea3405940e3c2d8276132dd062520f0c429591b"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.902014 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" event={"ID":"f71226fc-5b20-44c9-853f-adbed19887db","Type":"ContainerStarted","Data":"cbbdaf58419c61f0900c598477ddfd98fb99e78d3a2067f59338e71c97319867"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.903425 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.921312 4846 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nbllb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.921384 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" podUID="f71226fc-5b20-44c9-853f-adbed19887db" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.925454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:01 crc kubenswrapper[4846]: E0320 11:01:01.925933 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.425919896 +0000 UTC m=+233.792328132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.948936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bwgwb" event={"ID":"7b910eda-ae6a-4e2e-9ce2-9a139756b29f","Type":"ContainerStarted","Data":"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989"} Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.953810 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b7gtv" podStartSLOduration=8.95379058 podStartE2EDuration="8.95379058s" podCreationTimestamp="2026-03-20 11:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.952155926 +0000 UTC m=+233.318564162" watchObservedRunningTime="2026-03-20 11:01:01.95379058 +0000 UTC m=+233.320198816" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.955240 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-t4dz2" podStartSLOduration=181.955231969 podStartE2EDuration="3m1.955231969s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:01.87424756 +0000 UTC m=+233.240655796" watchObservedRunningTime="2026-03-20 11:01:01.955231969 +0000 UTC m=+233.321640205" Mar 20 11:01:01 crc kubenswrapper[4846]: I0320 11:01:01.989353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" event={"ID":"9f4e46fa-3163-4d53-b27c-bc04ba14576b","Type":"ContainerStarted","Data":"b1580d9c9a3aae854782df8b4edd14e8db6c8c2809e4392699db78e08a084f3f"} Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.007628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" event={"ID":"6d5e53b4-542d-4ab3-8248-dea9785709dd","Type":"ContainerStarted","Data":"8b6ad379150ea411d0aa8bd27113eac717737011614ed04b97cc8c6aeac313e9"} Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.008140 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.018312 4846 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-g4jbg container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.018382 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" podUID="6d5e53b4-542d-4ab3-8248-dea9785709dd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.026600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" event={"ID":"3a66972d-0296-4aee-8423-8455b65edb47","Type":"ContainerStarted","Data":"fa6fa1b1d0e76b6a73ec9438dbf2bea53c8b9a95e1130e2af4f942bac899bbad"} Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.027443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.032014 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.531989903 +0000 UTC m=+233.898398139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.045333 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:02 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:02 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:02 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.045418 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.052474 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" event={"ID":"fa67f6db-90b7-4eb4-ac3a-6e76785c92e2","Type":"ContainerStarted","Data":"569d3a72f47b361d54c1aff2e5d35418ea1b95ba822e374297ba37490b323673"} Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.055550 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-fkvwp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.055853 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fkvwp" podUID="ff4193a8-a16d-4803-b894-3c9d411677b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.060082 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5phhj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.060175 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.068764 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kbjmt" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.086530 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34098: no serving certificate available for the kubelet" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.114447 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" podStartSLOduration=183.114420842 podStartE2EDuration="3m3.114420842s" podCreationTimestamp="2026-03-20 10:57:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.104447042 +0000 UTC m=+233.470855278" watchObservedRunningTime="2026-03-20 11:01:02.114420842 +0000 UTC m=+233.480829078" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.141503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.142984 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.642959672 +0000 UTC m=+234.009367908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.164669 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-g7694" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.225341 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" podStartSLOduration=182.225320669 podStartE2EDuration="3m2.225320669s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.203605102 +0000 UTC m=+233.570013348" watchObservedRunningTime="2026-03-20 11:01:02.225320669 +0000 UTC m=+233.591728895" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.248753 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.249101 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.749078141 +0000 UTC m=+234.115486377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.249263 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.250451 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.750431607 +0000 UTC m=+234.116840033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.313506 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-pc5t4" podStartSLOduration=182.313483102 podStartE2EDuration="3m2.313483102s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.309325509 +0000 UTC m=+233.675733745" watchObservedRunningTime="2026-03-20 11:01:02.313483102 +0000 UTC m=+233.679891338" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.325293 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34112: no serving certificate available for the kubelet" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.353303 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.354077 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.854037038 +0000 UTC m=+234.220445274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.354271 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.354784 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.854776058 +0000 UTC m=+234.221184294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.420016 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wdb6b" podStartSLOduration=182.419966769 podStartE2EDuration="3m2.419966769s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.41702633 +0000 UTC m=+233.783434566" watchObservedRunningTime="2026-03-20 11:01:02.419966769 +0000 UTC m=+233.786375005" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.455688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.455949 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.955915471 +0000 UTC m=+234.322323707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.456052 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.456658 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:02.956645081 +0000 UTC m=+234.323053317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.557820 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.558195 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.058178705 +0000 UTC m=+234.424586941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.592382 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" podStartSLOduration=182.592361649 podStartE2EDuration="3m2.592361649s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.588451423 +0000 UTC m=+233.954859669" watchObservedRunningTime="2026-03-20 11:01:02.592361649 +0000 UTC m=+233.958769885" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.615451 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34126: no serving certificate available for the kubelet" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.659398 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.660071 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.160050829 +0000 UTC m=+234.526459065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.695206 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" podStartSLOduration=182.695185308 podStartE2EDuration="3m2.695185308s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.692738532 +0000 UTC m=+234.059146768" watchObservedRunningTime="2026-03-20 11:01:02.695185308 +0000 UTC m=+234.061593544" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.698059 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-b6szm" podStartSLOduration=182.698042116 podStartE2EDuration="3m2.698042116s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.641010424 +0000 UTC m=+234.007418660" watchObservedRunningTime="2026-03-20 11:01:02.698042116 +0000 UTC m=+234.064450352" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.733323 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34130: no serving certificate available for the kubelet" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.748595 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" podStartSLOduration=182.748560431 podStartE2EDuration="3m2.748560431s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.748286624 +0000 UTC m=+234.114694860" watchObservedRunningTime="2026-03-20 11:01:02.748560431 +0000 UTC m=+234.114968667" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.760054 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.760485 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.260470652 +0000 UTC m=+234.626878888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.794152 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" podStartSLOduration=182.794125842 podStartE2EDuration="3m2.794125842s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.790117763 +0000 UTC m=+234.156525999" watchObservedRunningTime="2026-03-20 11:01:02.794125842 +0000 UTC m=+234.160534078" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.847616 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34136: no serving certificate available for the kubelet" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.855452 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fmcbh" podStartSLOduration=8.855431839 podStartE2EDuration="8.855431839s" podCreationTimestamp="2026-03-20 11:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.853080775 +0000 UTC m=+234.219489001" watchObservedRunningTime="2026-03-20 11:01:02.855431839 +0000 UTC m=+234.221840075" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.861380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.861815 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.361801032 +0000 UTC m=+234.728209268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.877190 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sn4x4" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.919517 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-bwgwb" podStartSLOduration=182.919499751 podStartE2EDuration="3m2.919499751s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.918026761 +0000 UTC m=+234.284434997" watchObservedRunningTime="2026-03-20 11:01:02.919499751 +0000 UTC m=+234.285907987" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.957247 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mdwxk" podStartSLOduration=182.957226581 podStartE2EDuration="3m2.957226581s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:02.948323419 +0000 UTC m=+234.314731655" watchObservedRunningTime="2026-03-20 11:01:02.957226581 +0000 UTC m=+234.323634817" Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.962452 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:02 crc kubenswrapper[4846]: E0320 11:01:02.962708 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.462674128 +0000 UTC m=+234.829082364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:02 crc kubenswrapper[4846]: I0320 11:01:02.964643 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34152: no serving certificate available for the kubelet" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.040165 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:03 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:03 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:03 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.040648 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.070857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.071289 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.571271223 +0000 UTC m=+234.937679459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.087372 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34156: no serving certificate available for the kubelet" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.092950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7zwsq" event={"ID":"437d4e9e-f459-4b74-8182-58aa5f25df5e","Type":"ContainerStarted","Data":"d9f23e6a7f74aa34663663e6a670be5bd3e4ecb1f19cbfeb08dff2f2d399c0f0"} Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.093836 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7zwsq" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.101274 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7n8h2" podStartSLOduration=183.101253273 podStartE2EDuration="3m3.101253273s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:03.043275256 +0000 UTC m=+234.409683492" watchObservedRunningTime="2026-03-20 11:01:03.101253273 +0000 UTC m=+234.467661509" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.135558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" event={"ID":"17ac0e08-c160-4a84-9f64-7ecef5dd90d4","Type":"ContainerStarted","Data":"ca795fdd69f669ee2e11dc573d943dc83afee0632658f8bc2e2dd7013104b6de"} Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.165141 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" event={"ID":"ad7112f5-c4c3-4a89-9b97-18e3caea5913","Type":"ContainerStarted","Data":"4c6bb2ce4a21cdd86924dd3036dec70c5b12332cc10a26c88f3c3fa9093c90a7"} Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.171654 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.172146 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.672131469 +0000 UTC m=+235.038539695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.186253 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-g4jbg" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.197812 4846 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wpkrb container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.197882 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" podUID="b884ddf6-9cfc-41d7-9659-dd3c86dc0337" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.222486 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34158: no serving certificate available for the kubelet" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.229892 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6q4br" podStartSLOduration=183.22987304 podStartE2EDuration="3m3.22987304s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:03.222372417 +0000 UTC m=+234.588780663" watchObservedRunningTime="2026-03-20 11:01:03.22987304 +0000 UTC m=+234.596281276" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.230625 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sv5bj" podStartSLOduration=183.23062003 podStartE2EDuration="3m3.23062003s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:03.156844935 +0000 UTC m=+234.523253171" watchObservedRunningTime="2026-03-20 11:01:03.23062003 +0000 UTC m=+234.597028256" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.286770 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.291225 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.291347 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.296745 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.796712336 +0000 UTC m=+235.163120572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.305990 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.398432 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.399741 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.89970814 +0000 UTC m=+235.266116376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.400108 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.400522 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:03.900514041 +0000 UTC m=+235.266922277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.406046 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.407038 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.407130 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.415380 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.415616 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.442278 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7zwsq" podStartSLOduration=10.442256889 podStartE2EDuration="10.442256889s" podCreationTimestamp="2026-03-20 11:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:03.427839 +0000 UTC m=+234.794247236" watchObservedRunningTime="2026-03-20 11:01:03.442256889 +0000 UTC m=+234.808665125" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.504339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.504758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.504810 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.004773389 +0000 UTC m=+235.371181625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.504862 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.504986 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.505359 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.005352665 +0000 UTC m=+235.371760891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.605780 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.606147 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.606192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.606439 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.106418616 +0000 UTC m=+235.472826852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.606589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.638324 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.708379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.708854 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.208841655 +0000 UTC m=+235.575249891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.735023 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.739862 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.741076 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.746818 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.751630 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.777110 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.810494 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.810600 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.310579764 +0000 UTC m=+235.676988000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.810915 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.810953 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.810996 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.811027 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s4jw\" (UniqueName: \"kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.811340 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.311331825 +0000 UTC m=+235.677740061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.911647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.911825 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.4117889 +0000 UTC m=+235.778197136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.912035 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.912084 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.912136 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.912167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s4jw\" (UniqueName: \"kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.912926 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: E0320 11:01:03.913223 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.413208019 +0000 UTC m=+235.779616255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.913624 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.931307 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34172: no serving certificate available for the kubelet" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.955689 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.957776 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.961083 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.963206 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s4jw\" (UniqueName: \"kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw\") pod \"community-operators-56jhc\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:03 crc kubenswrapper[4846]: I0320 11:01:03.965299 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.013041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.013492 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.513474718 +0000 UTC m=+235.879882954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.032543 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:04 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:04 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:04 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.032604 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.052170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nbllb" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.061289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.117737 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.118268 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.118292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj6gz\" (UniqueName: \"kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.118322 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.118726 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.618712182 +0000 UTC m=+235.985120418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.148581 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.149654 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.208379 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.227030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.230625 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.730595096 +0000 UTC m=+236.097003392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.230779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.230854 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.230914 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj6gz\" (UniqueName: \"kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.230988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgwxs\" (UniqueName: \"kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.231034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.231114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.231724 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.232404 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.253556 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" containerName="controller-manager" containerID="cri-o://46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529" gracePeriod=30 Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.268366 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wpkrb" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.279700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj6gz\" (UniqueName: \"kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz\") pod \"certified-operators-tt9fr\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.332434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgwxs\" (UniqueName: \"kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.332790 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.332966 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.333119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.337213 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.837193637 +0000 UTC m=+236.203601873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.337655 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.363262 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.367199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.382709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.419499 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgwxs\" (UniqueName: \"kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs\") pod \"community-operators-lh8ff\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.435700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.436598 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.437227 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:04.93717814 +0000 UTC m=+236.303586376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.437430 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.437542 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.437634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dfm\" (UniqueName: \"kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.469112 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.538857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.538941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.538984 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dfm\" (UniqueName: \"kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.539026 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.539396 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.039379723 +0000 UTC m=+236.405787959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.540497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.540730 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.555755 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.568168 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dfm\" (UniqueName: \"kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm\") pod \"certified-operators-4z7gv\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.592414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.645620 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.645997 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.145972753 +0000 UTC m=+236.512380989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.695715 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:01:04 crc kubenswrapper[4846]: W0320 11:01:04.705985 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b17dfdf_1984_4cde_958e_021958ad8e0c.slice/crio-faa3a5a937d1a11e52535d180ca9b905119494c913197e43b6867d65abeff89f WatchSource:0}: Error finding container faa3a5a937d1a11e52535d180ca9b905119494c913197e43b6867d65abeff89f: Status 404 returned error can't find the container with id faa3a5a937d1a11e52535d180ca9b905119494c913197e43b6867d65abeff89f Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.737326 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.747213 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.747694 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.247676223 +0000 UTC m=+236.614084459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.848065 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.848311 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.348276401 +0000 UTC m=+236.714684637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.848528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.848996 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.34897813 +0000 UTC m=+236.715386366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.951973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.952540 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.452513799 +0000 UTC m=+236.818922035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.952860 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:04 crc kubenswrapper[4846]: E0320 11:01:04.953264 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.453257418 +0000 UTC m=+236.819665654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:04 crc kubenswrapper[4846]: I0320 11:01:04.979816 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.023190 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57f9c6d465-rxwtm"] Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.023477 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" containerName="controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.023491 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" containerName="controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.023604 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" containerName="controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.025093 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.054402 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qff8q\" (UniqueName: \"kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q\") pod \"90413730-bfef-4d28-9c45-7aa031e5af17\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.054776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.054808 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca\") pod \"90413730-bfef-4d28-9c45-7aa031e5af17\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.054876 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config\") pod \"90413730-bfef-4d28-9c45-7aa031e5af17\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.054983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert\") pod \"90413730-bfef-4d28-9c45-7aa031e5af17\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055072 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles\") pod \"90413730-bfef-4d28-9c45-7aa031e5af17\" (UID: \"90413730-bfef-4d28-9c45-7aa031e5af17\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055291 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055322 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6lbn\" (UniqueName: \"kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055339 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055358 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.055393 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.056634 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57f9c6d465-rxwtm"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.058028 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca" (OuterVolumeSpecName: "client-ca") pod "90413730-bfef-4d28-9c45-7aa031e5af17" (UID: "90413730-bfef-4d28-9c45-7aa031e5af17"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.058058 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "90413730-bfef-4d28-9c45-7aa031e5af17" (UID: "90413730-bfef-4d28-9c45-7aa031e5af17"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.058143 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.558122903 +0000 UTC m=+236.924531139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.058765 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config" (OuterVolumeSpecName: "config") pod "90413730-bfef-4d28-9c45-7aa031e5af17" (UID: "90413730-bfef-4d28-9c45-7aa031e5af17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.065676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q" (OuterVolumeSpecName: "kube-api-access-qff8q") pod "90413730-bfef-4d28-9c45-7aa031e5af17" (UID: "90413730-bfef-4d28-9c45-7aa031e5af17"). InnerVolumeSpecName "kube-api-access-qff8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.067027 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "90413730-bfef-4d28-9c45-7aa031e5af17" (UID: "90413730-bfef-4d28-9c45-7aa031e5af17"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.077494 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:05 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:05 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:05 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.077544 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.147388 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57f9c6d465-rxwtm"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156542 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6lbn\" (UniqueName: \"kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156597 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156630 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156797 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156808 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qff8q\" (UniqueName: \"kubernetes.io/projected/90413730-bfef-4d28-9c45-7aa031e5af17-kube-api-access-qff8q\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156818 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156827 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90413730-bfef-4d28-9c45-7aa031e5af17-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.156838 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90413730-bfef-4d28-9c45-7aa031e5af17-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.158072 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-h6lbn proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" podUID="4a076431-95b2-44f2-86c2-3f029ad41b26" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.159145 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.159868 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.659846002 +0000 UTC m=+237.026254238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.160047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.160743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.182016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.200108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6lbn\" (UniqueName: \"kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn\") pod \"controller-manager-57f9c6d465-rxwtm\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.229412 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.260721 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.261174 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.761132439 +0000 UTC m=+237.127540675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.273044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.273584 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.773553335 +0000 UTC m=+237.139961731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.274662 4846 ???:1] "http: TLS handshake error from 192.168.126.11:34176: no serving certificate available for the kubelet" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.276058 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.311388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" event={"ID":"ad7112f5-c4c3-4a89-9b97-18e3caea5913","Type":"ContainerStarted","Data":"1882793c9ef93aa4ce58af7652bed726e8b1ed3e4b6b1289977759b0c7fa9fa6"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.313872 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9264ef84-1347-4297-b243-6c17db904b30","Type":"ContainerStarted","Data":"76318b8d8ef3cda4a8178f22741e230fc0c210abd92ccbbbb851d0caa773e126"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.313962 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9264ef84-1347-4297-b243-6c17db904b30","Type":"ContainerStarted","Data":"fa2fff14128ea48a92baec5335f269453c787cae68d480a979d0a81aabdbd701"} Mar 20 11:01:05 crc kubenswrapper[4846]: W0320 11:01:05.326975 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod408ee9c3_1b20_4002_850f_328b27b972d0.slice/crio-8105e2e50c571fe2eb64207670da80cd57289abd8a3a2f5cf97fa65fa323adb7 WatchSource:0}: Error finding container 8105e2e50c571fe2eb64207670da80cd57289abd8a3a2f5cf97fa65fa323adb7: Status 404 returned error can't find the container with id 8105e2e50c571fe2eb64207670da80cd57289abd8a3a2f5cf97fa65fa323adb7 Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.327961 4846 generic.go:334] "Generic (PLEG): container finished" podID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerID="d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9" exitCode=0 Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.337153 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.337115104 podStartE2EDuration="2.337115104s" podCreationTimestamp="2026-03-20 11:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:05.332534829 +0000 UTC m=+236.698943105" watchObservedRunningTime="2026-03-20 11:01:05.337115104 +0000 UTC m=+236.703523340" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.340037 4846 generic.go:334] "Generic (PLEG): container finished" podID="90413730-bfef-4d28-9c45-7aa031e5af17" containerID="46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529" exitCode=0 Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.340979 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.341088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerDied","Data":"d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.341125 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerStarted","Data":"faa3a5a937d1a11e52535d180ca9b905119494c913197e43b6867d65abeff89f"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.341137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" event={"ID":"90413730-bfef-4d28-9c45-7aa031e5af17","Type":"ContainerDied","Data":"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.341151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qwzz" event={"ID":"90413730-bfef-4d28-9c45-7aa031e5af17","Type":"ContainerDied","Data":"b704a5db648bc3de6c872bb5a22cfc70130bff4863813e921cef354ac7650ecd"} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.341175 4846 scope.go:117] "RemoveContainer" containerID="46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.344796 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.347212 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerName="route-controller-manager" containerID="cri-o://dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11" gracePeriod=30 Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.362802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.374799 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.376036 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.876011334 +0000 UTC m=+237.242419570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.388425 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.394112 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.397848 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qwzz"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.435048 4846 scope.go:117] "RemoveContainer" containerID="46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.438076 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529\": container with ID starting with 46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529 not found: ID does not exist" containerID="46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.438135 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529"} err="failed to get container status \"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529\": rpc error: code = NotFound desc = could not find container \"46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529\": container with ID starting with 46b81afcfc9bbceccb1f98c5a8ab15711b895bf8309ca594d7fe2f62b8f3a529 not found: ID does not exist" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.477464 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert\") pod \"4a076431-95b2-44f2-86c2-3f029ad41b26\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.477554 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config\") pod \"4a076431-95b2-44f2-86c2-3f029ad41b26\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.477623 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles\") pod \"4a076431-95b2-44f2-86c2-3f029ad41b26\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.478812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca\") pod \"4a076431-95b2-44f2-86c2-3f029ad41b26\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.478915 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6lbn\" (UniqueName: \"kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn\") pod \"4a076431-95b2-44f2-86c2-3f029ad41b26\" (UID: \"4a076431-95b2-44f2-86c2-3f029ad41b26\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.479218 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.479538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca" (OuterVolumeSpecName: "client-ca") pod "4a076431-95b2-44f2-86c2-3f029ad41b26" (UID: "4a076431-95b2-44f2-86c2-3f029ad41b26"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.479635 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config" (OuterVolumeSpecName: "config") pod "4a076431-95b2-44f2-86c2-3f029ad41b26" (UID: "4a076431-95b2-44f2-86c2-3f029ad41b26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.479656 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:05.979640095 +0000 UTC m=+237.346048401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.479723 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4a076431-95b2-44f2-86c2-3f029ad41b26" (UID: "4a076431-95b2-44f2-86c2-3f029ad41b26"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.507847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4a076431-95b2-44f2-86c2-3f029ad41b26" (UID: "4a076431-95b2-44f2-86c2-3f029ad41b26"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.508270 4846 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.508572 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn" (OuterVolumeSpecName: "kube-api-access-h6lbn") pod "4a076431-95b2-44f2-86c2-3f029ad41b26" (UID: "4a076431-95b2-44f2-86c2-3f029ad41b26"). InnerVolumeSpecName "kube-api-access-h6lbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.580636 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.580846 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 11:01:06.08081334 +0000 UTC m=+237.447221586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581231 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6lbn\" (UniqueName: \"kubernetes.io/projected/4a076431-95b2-44f2-86c2-3f029ad41b26-kube-api-access-h6lbn\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581259 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a076431-95b2-44f2-86c2-3f029ad41b26-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581280 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581298 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.581309 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a076431-95b2-44f2-86c2-3f029ad41b26-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.581509 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 11:01:06.081497788 +0000 UTC m=+237.447906234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cn4z4" (UID: "a9392c91-350e-4052-a357-5c4da7651485") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.593602 4846 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-20T11:01:05.508354912Z","Handler":null,"Name":""} Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.602782 4846 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.602841 4846 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.682390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.689722 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.771722 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.771772 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.779521 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.783860 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.795284 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.795345 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.839261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cn4z4\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.912048 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.948221 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:01:05 crc kubenswrapper[4846]: E0320 11:01:05.948624 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerName="route-controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.948647 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerName="route-controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.948755 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerName="route-controller-manager" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.949687 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.952427 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.958820 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.987323 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca\") pod \"ac8d3b56-111a-440c-a3c3-9ff9708af766\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.987425 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47wrn\" (UniqueName: \"kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn\") pod \"ac8d3b56-111a-440c-a3c3-9ff9708af766\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.987551 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert\") pod \"ac8d3b56-111a-440c-a3c3-9ff9708af766\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.987749 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config\") pod \"ac8d3b56-111a-440c-a3c3-9ff9708af766\" (UID: \"ac8d3b56-111a-440c-a3c3-9ff9708af766\") " Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.988148 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.988191 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p45cb\" (UniqueName: \"kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.988321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.989401 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca" (OuterVolumeSpecName: "client-ca") pod "ac8d3b56-111a-440c-a3c3-9ff9708af766" (UID: "ac8d3b56-111a-440c-a3c3-9ff9708af766"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.989505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config" (OuterVolumeSpecName: "config") pod "ac8d3b56-111a-440c-a3c3-9ff9708af766" (UID: "ac8d3b56-111a-440c-a3c3-9ff9708af766"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:05 crc kubenswrapper[4846]: I0320 11:01:05.999424 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ac8d3b56-111a-440c-a3c3-9ff9708af766" (UID: "ac8d3b56-111a-440c-a3c3-9ff9708af766"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.001967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn" (OuterVolumeSpecName: "kube-api-access-47wrn") pod "ac8d3b56-111a-440c-a3c3-9ff9708af766" (UID: "ac8d3b56-111a-440c-a3c3-9ff9708af766"). InnerVolumeSpecName "kube-api-access-47wrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.040699 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:06 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:06 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:06 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.050773 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.067832 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.090982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091102 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p45cb\" (UniqueName: \"kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091178 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091192 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac8d3b56-111a-440c-a3c3-9ff9708af766-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091203 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47wrn\" (UniqueName: \"kubernetes.io/projected/ac8d3b56-111a-440c-a3c3-9ff9708af766-kube-api-access-47wrn\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.091217 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8d3b56-111a-440c-a3c3-9ff9708af766-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.092236 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.092329 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.112722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p45cb\" (UniqueName: \"kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb\") pod \"redhat-marketplace-6vwj9\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.271706 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.344769 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.345957 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.349626 4846 generic.go:334] "Generic (PLEG): container finished" podID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerID="27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.349728 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerDied","Data":"27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.349752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerStarted","Data":"524dbe9258cd5b4f00bb342c6667ea970055de8fab93d7b0473e9ab8dfac6a7b"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.356621 4846 generic.go:334] "Generic (PLEG): container finished" podID="408ee9c3-1b20-4002-850f-328b27b972d0" containerID="8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.356753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerDied","Data":"8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.356805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerStarted","Data":"8105e2e50c571fe2eb64207670da80cd57289abd8a3a2f5cf97fa65fa323adb7"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.364333 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.371872 4846 generic.go:334] "Generic (PLEG): container finished" podID="53ec1a91-8224-4f4e-a4ed-d0ee907b0354" containerID="5203b292d231b364464cd68a9277c86c387901b379f9a0fde9581e81195e5c86" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.371991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" event={"ID":"53ec1a91-8224-4f4e-a4ed-d0ee907b0354","Type":"ContainerDied","Data":"5203b292d231b364464cd68a9277c86c387901b379f9a0fde9581e81195e5c86"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.374353 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-fkvwp container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.374482 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fkvwp" podUID="ff4193a8-a16d-4803-b894-3c9d411677b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.374982 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-fkvwp container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.375037 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fkvwp" podUID="ff4193a8-a16d-4803-b894-3c9d411677b2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.385232 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ba6d606-56e4-4285-887a-4b86f8198046" containerID="2b6a518874b790be561c435c3114c0df984e1e56213da3d0baffc76761dc1335" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.385343 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerDied","Data":"2b6a518874b790be561c435c3114c0df984e1e56213da3d0baffc76761dc1335"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.385373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerStarted","Data":"75018c7f04688ad7561a3f230f6bbef978b7678fb85c43ca20f8bfc3e8dd6ae0"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.397365 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.397518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mq5x\" (UniqueName: \"kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.397614 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.408283 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.409516 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.415004 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.415374 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.420049 4846 generic.go:334] "Generic (PLEG): container finished" podID="ac8d3b56-111a-440c-a3c3-9ff9708af766" containerID="dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.420283 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.420347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" event={"ID":"ac8d3b56-111a-440c-a3c3-9ff9708af766","Type":"ContainerDied","Data":"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.420391 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7" event={"ID":"ac8d3b56-111a-440c-a3c3-9ff9708af766","Type":"ContainerDied","Data":"a99b79b89fc3031ff4e601c03e7e95714d8ec971a7d7f318a53a4467048b60ad"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.420413 4846 scope.go:117] "RemoveContainer" containerID="dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.426374 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.428605 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.484308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" event={"ID":"ad7112f5-c4c3-4a89-9b97-18e3caea5913","Type":"ContainerStarted","Data":"4ba5de7e03ae7d4a86f5078d84284eaec38359ba5fda96598e603d47f1a6d4e4"} Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.503482 4846 scope.go:117] "RemoveContainer" containerID="dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.503992 4846 generic.go:334] "Generic (PLEG): container finished" podID="9264ef84-1347-4297-b243-6c17db904b30" containerID="76318b8d8ef3cda4a8178f22741e230fc0c210abd92ccbbbb851d0caa773e126" exitCode=0 Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.504172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f9c6d465-rxwtm" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.504303 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9264ef84-1347-4297-b243-6c17db904b30","Type":"ContainerDied","Data":"76318b8d8ef3cda4a8178f22741e230fc0c210abd92ccbbbb851d0caa773e126"} Mar 20 11:01:06 crc kubenswrapper[4846]: E0320 11:01:06.507572 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11\": container with ID starting with dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11 not found: ID does not exist" containerID="dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.507649 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11"} err="failed to get container status \"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11\": rpc error: code = NotFound desc = could not find container \"dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11\": container with ID starting with dc0a491bf7b930e19ab34746af8ac0fbedd75f613118ba1b99fd298284da7f11 not found: ID does not exist" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.508362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.508720 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.508993 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.509065 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mq5x\" (UniqueName: \"kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.509140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.509175 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.509420 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.520501 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sjmnj" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.562579 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mq5x\" (UniqueName: \"kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x\") pod \"redhat-marketplace-6qggw\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.563210 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.602282 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5lnm7"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.613072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.613207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.614975 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.658641 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57f9c6d465-rxwtm"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.665775 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-57f9c6d465-rxwtm"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.682961 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.683414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.784387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.788486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.807770 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.940194 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.948033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.953825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:01:06 crc kubenswrapper[4846]: I0320 11:01:06.955091 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.027727 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjvlg\" (UniqueName: \"kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.027777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.027831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.027992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.048306 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:07 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:07 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:07 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.048377 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.061168 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.062023 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.063106 4846 patch_prober.go:28] interesting pod/console-f9d7485db-bwgwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.063156 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bwgwb" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.128847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.129296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjvlg\" (UniqueName: \"kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.129335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.129744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.130200 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.159844 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.165139 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjvlg\" (UniqueName: \"kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg\") pod \"redhat-operators-kh2d5\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: W0320 11:01:07.211366 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod131db4bd_a795_401c_a9e7_b66f4512dc32.slice/crio-7416ad4f63b2f88eecdd33df72960cd715dd51b0f92b9bf9f51d4f8be9ed7f52 WatchSource:0}: Error finding container 7416ad4f63b2f88eecdd33df72960cd715dd51b0f92b9bf9f51d4f8be9ed7f52: Status 404 returned error can't find the container with id 7416ad4f63b2f88eecdd33df72960cd715dd51b0f92b9bf9f51d4f8be9ed7f52 Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.307307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.312100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.356245 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a076431-95b2-44f2-86c2-3f029ad41b26" path="/var/lib/kubelet/pods/4a076431-95b2-44f2-86c2-3f029ad41b26/volumes" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.357631 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.358543 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90413730-bfef-4d28-9c45-7aa031e5af17" path="/var/lib/kubelet/pods/90413730-bfef-4d28-9c45-7aa031e5af17/volumes" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.359186 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac8d3b56-111a-440c-a3c3-9ff9708af766" path="/var/lib/kubelet/pods/ac8d3b56-111a-440c-a3c3-9ff9708af766/volumes" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.360515 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.364410 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.364538 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.436679 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.441319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9pfz\" (UniqueName: \"kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.441414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.441460 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.441832 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.446870 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.447439 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.447708 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.448029 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.448229 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.449058 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.466525 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.547707 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.547936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548013 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548045 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5qms\" (UniqueName: \"kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548082 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548276 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548343 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9pfz\" (UniqueName: \"kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.548931 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.550367 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.574348 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerID="018ae237d92648099fca8b5006a1637688b0e36b99650e00d7aab8ead8d83434" exitCode=0 Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.574571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerDied","Data":"018ae237d92648099fca8b5006a1637688b0e36b99650e00d7aab8ead8d83434"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.574640 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerStarted","Data":"1e89c210582146a3985d6ff388bba58b9bc4a545217d0730ee1cef8c249e1e37"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.580199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9pfz\" (UniqueName: \"kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz\") pod \"redhat-operators-nhrbq\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.583228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e32b5559-d9ec-4b43-ba99-c80678cf5e41","Type":"ContainerStarted","Data":"4c1dd2c480593c5e8701840f7148529e8e9ba240164edb908c3f12bb55519b1e"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.587545 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerStarted","Data":"7416ad4f63b2f88eecdd33df72960cd715dd51b0f92b9bf9f51d4f8be9ed7f52"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.598290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" event={"ID":"ad7112f5-c4c3-4a89-9b97-18e3caea5913","Type":"ContainerStarted","Data":"60dfcf48a8878dcb323193a48b28c3658e1af722e6c2e88e07e4e6e5c342234c"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.615485 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" event={"ID":"a9392c91-350e-4052-a357-5c4da7651485","Type":"ContainerStarted","Data":"c0d3b3a8cdab4ab5bd7410f5ee6afdd8a0df9d340cf87802b2110240959c0602"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.615551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" event={"ID":"a9392c91-350e-4052-a357-5c4da7651485","Type":"ContainerStarted","Data":"380f0f7bc5484ec624cc8d631e7da172a1ac3fb95607581cceaec1f6ec41a089"} Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.615673 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.645725 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gsdtt" podStartSLOduration=14.645691308 podStartE2EDuration="14.645691308s" podCreationTimestamp="2026-03-20 11:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:07.634002053 +0000 UTC m=+239.000410309" watchObservedRunningTime="2026-03-20 11:01:07.645691308 +0000 UTC m=+239.012099544" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.649673 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.649744 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5qms\" (UniqueName: \"kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.649831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.649928 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.651074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.651567 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.663491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.664290 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" podStartSLOduration=187.664261541 podStartE2EDuration="3m7.664261541s" podCreationTimestamp="2026-03-20 10:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:07.663674125 +0000 UTC m=+239.030082371" watchObservedRunningTime="2026-03-20 11:01:07.664261541 +0000 UTC m=+239.030669777" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.678268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5qms\" (UniqueName: \"kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms\") pod \"route-controller-manager-67776fb75d-ntnms\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.794483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.811696 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.894591 4846 ???:1] "http: TLS handshake error from 192.168.126.11:36486: no serving certificate available for the kubelet" Mar 20 11:01:07 crc kubenswrapper[4846]: I0320 11:01:07.898138 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.044824 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:08 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:08 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:08 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.044958 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.170217 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.217142 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.261333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume\") pod \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.261845 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvk9l\" (UniqueName: \"kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l\") pod \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.261949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume\") pod \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\" (UID: \"53ec1a91-8224-4f4e-a4ed-d0ee907b0354\") " Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.261987 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir\") pod \"9264ef84-1347-4297-b243-6c17db904b30\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.262045 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access\") pod \"9264ef84-1347-4297-b243-6c17db904b30\" (UID: \"9264ef84-1347-4297-b243-6c17db904b30\") " Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.263687 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9264ef84-1347-4297-b243-6c17db904b30" (UID: "9264ef84-1347-4297-b243-6c17db904b30"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.264025 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume" (OuterVolumeSpecName: "config-volume") pod "53ec1a91-8224-4f4e-a4ed-d0ee907b0354" (UID: "53ec1a91-8224-4f4e-a4ed-d0ee907b0354"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.276838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9264ef84-1347-4297-b243-6c17db904b30" (UID: "9264ef84-1347-4297-b243-6c17db904b30"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.277513 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "53ec1a91-8224-4f4e-a4ed-d0ee907b0354" (UID: "53ec1a91-8224-4f4e-a4ed-d0ee907b0354"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.278134 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l" (OuterVolumeSpecName: "kube-api-access-kvk9l") pod "53ec1a91-8224-4f4e-a4ed-d0ee907b0354" (UID: "53ec1a91-8224-4f4e-a4ed-d0ee907b0354"). InnerVolumeSpecName "kube-api-access-kvk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.364254 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.364292 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvk9l\" (UniqueName: \"kubernetes.io/projected/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-kube-api-access-kvk9l\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.364302 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53ec1a91-8224-4f4e-a4ed-d0ee907b0354-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.364311 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9264ef84-1347-4297-b243-6c17db904b30-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.364320 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9264ef84-1347-4297-b243-6c17db904b30-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.429762 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.456214 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.628397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" event={"ID":"5806af56-cc91-46ec-9255-034a44f08001","Type":"ContainerStarted","Data":"53ad959f044db20eef968c12b3af3f0f57966f3854fd792f9cd12f59c1a9f9cb"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.631666 4846 generic.go:334] "Generic (PLEG): container finished" podID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerID="217dee6b0998b8db43f8127c84e961c5451fee355873d531123fc419c35d02a1" exitCode=0 Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.631769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerDied","Data":"217dee6b0998b8db43f8127c84e961c5451fee355873d531123fc419c35d02a1"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.631853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerStarted","Data":"a189629dffd267da69a8916dcc27fe031e6418fbaa757b896e8df4a12c75e77e"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.636182 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.636216 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9264ef84-1347-4297-b243-6c17db904b30","Type":"ContainerDied","Data":"fa2fff14128ea48a92baec5335f269453c787cae68d480a979d0a81aabdbd701"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.636250 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa2fff14128ea48a92baec5335f269453c787cae68d480a979d0a81aabdbd701" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.638082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerStarted","Data":"120c84aaf198f9cc55ea03048a64e1603bae415f39dc2e7bfe920ce69e5e0ed9"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.648749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" event={"ID":"53ec1a91-8224-4f4e-a4ed-d0ee907b0354","Type":"ContainerDied","Data":"05fae997e711156d655a2f7612c92b2e5c2bb6414cbf70f2ce52d3cdfe75410c"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.649163 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05fae997e711156d655a2f7612c92b2e5c2bb6414cbf70f2ce52d3cdfe75410c" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.649261 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z" Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.661590 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e32b5559-d9ec-4b43-ba99-c80678cf5e41","Type":"ContainerStarted","Data":"8622371f9ea9977084eb2e19da50014b0da628f84650e91115dcc262243f2b5a"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.670663 4846 generic.go:334] "Generic (PLEG): container finished" podID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerID="6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834" exitCode=0 Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.671031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerDied","Data":"6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834"} Mar 20 11:01:08 crc kubenswrapper[4846]: I0320 11:01:08.683693 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.683663423 podStartE2EDuration="2.683663423s" podCreationTimestamp="2026-03-20 11:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:08.68135784 +0000 UTC m=+240.047766076" watchObservedRunningTime="2026-03-20 11:01:08.683663423 +0000 UTC m=+240.050071839" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.033036 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:09 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Mar 20 11:01:09 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:09 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.033111 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.433525 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:09 crc kubenswrapper[4846]: E0320 11:01:09.435675 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec1a91-8224-4f4e-a4ed-d0ee907b0354" containerName="collect-profiles" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.435695 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec1a91-8224-4f4e-a4ed-d0ee907b0354" containerName="collect-profiles" Mar 20 11:01:09 crc kubenswrapper[4846]: E0320 11:01:09.435708 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9264ef84-1347-4297-b243-6c17db904b30" containerName="pruner" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.435715 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9264ef84-1347-4297-b243-6c17db904b30" containerName="pruner" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.435838 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ec1a91-8224-4f4e-a4ed-d0ee907b0354" containerName="collect-profiles" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.435855 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9264ef84-1347-4297-b243-6c17db904b30" containerName="pruner" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.436324 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.448630 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.449026 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.455623 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.455876 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.456346 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.456487 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.456602 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.460351 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.487000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdgw\" (UniqueName: \"kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.487063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.487091 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.487118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.487158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.592151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.592283 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdgw\" (UniqueName: \"kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.592326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.592354 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.592454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.596827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.596888 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.597188 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.606996 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.612770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdgw\" (UniqueName: \"kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw\") pod \"controller-manager-7dddc88b74-pdsbg\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.678550 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.678610 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.685294 4846 generic.go:334] "Generic (PLEG): container finished" podID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerID="78310e72e3d3c96b337ca94613a7e79f387a9afa753caafcbf85328ea54a708d" exitCode=0 Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.685369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerDied","Data":"78310e72e3d3c96b337ca94613a7e79f387a9afa753caafcbf85328ea54a708d"} Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.689342 4846 generic.go:334] "Generic (PLEG): container finished" podID="e32b5559-d9ec-4b43-ba99-c80678cf5e41" containerID="8622371f9ea9977084eb2e19da50014b0da628f84650e91115dcc262243f2b5a" exitCode=0 Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.689397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e32b5559-d9ec-4b43-ba99-c80678cf5e41","Type":"ContainerDied","Data":"8622371f9ea9977084eb2e19da50014b0da628f84650e91115dcc262243f2b5a"} Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.694660 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" event={"ID":"5806af56-cc91-46ec-9255-034a44f08001","Type":"ContainerStarted","Data":"57961d873c7cc9b9911d43cdf0f3d121f5d6b5e8aca8069865b06e8d2234d07b"} Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.695299 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.737014 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" podStartSLOduration=6.736993782 podStartE2EDuration="6.736993782s" podCreationTimestamp="2026-03-20 11:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:09.735576924 +0000 UTC m=+241.101985160" watchObservedRunningTime="2026-03-20 11:01:09.736993782 +0000 UTC m=+241.103402008" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.765608 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:09 crc kubenswrapper[4846]: I0320 11:01:09.812035 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.038055 4846 patch_prober.go:28] interesting pod/router-default-5444994796-s2rfb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 11:01:10 crc kubenswrapper[4846]: [+]has-synced ok Mar 20 11:01:10 crc kubenswrapper[4846]: [+]process-running ok Mar 20 11:01:10 crc kubenswrapper[4846]: healthz check failed Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.038174 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-s2rfb" podUID="1d80adab-bdc5-4c87-88af-c698df414d0f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.054290 4846 ???:1] "http: TLS handshake error from 192.168.126.11:36488: no serving certificate available for the kubelet" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.150030 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:10 crc kubenswrapper[4846]: W0320 11:01:10.171189 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8858bf7_b571_4de3_9bc3_82e527e9b8b5.slice/crio-e361f54d025ed569e5befd62b7e405f71f18ad826fdf3a6b5c67989e60020ba4 WatchSource:0}: Error finding container e361f54d025ed569e5befd62b7e405f71f18ad826fdf3a6b5c67989e60020ba4: Status 404 returned error can't find the container with id e361f54d025ed569e5befd62b7e405f71f18ad826fdf3a6b5c67989e60020ba4 Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.211531 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.725724 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" event={"ID":"a8858bf7-b571-4de3-9bc3-82e527e9b8b5","Type":"ContainerStarted","Data":"76689621b1edcf8e9f0a435c9a8953aed400570db3e2667fd1f3f5a1c2459fdb"} Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.725784 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.725794 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" event={"ID":"a8858bf7-b571-4de3-9bc3-82e527e9b8b5","Type":"ContainerStarted","Data":"e361f54d025ed569e5befd62b7e405f71f18ad826fdf3a6b5c67989e60020ba4"} Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.736159 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:10 crc kubenswrapper[4846]: I0320 11:01:10.762137 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" podStartSLOduration=5.762109078 podStartE2EDuration="5.762109078s" podCreationTimestamp="2026-03-20 11:01:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:10.756547498 +0000 UTC m=+242.122955754" watchObservedRunningTime="2026-03-20 11:01:10.762109078 +0000 UTC m=+242.128517314" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.049101 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.055973 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-s2rfb" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.354166 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.536390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir\") pod \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.536824 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access\") pod \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\" (UID: \"e32b5559-d9ec-4b43-ba99-c80678cf5e41\") " Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.536914 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e32b5559-d9ec-4b43-ba99-c80678cf5e41" (UID: "e32b5559-d9ec-4b43-ba99-c80678cf5e41"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.537561 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.549196 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e32b5559-d9ec-4b43-ba99-c80678cf5e41" (UID: "e32b5559-d9ec-4b43-ba99-c80678cf5e41"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.639057 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e32b5559-d9ec-4b43-ba99-c80678cf5e41-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.764884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e32b5559-d9ec-4b43-ba99-c80678cf5e41","Type":"ContainerDied","Data":"4c1dd2c480593c5e8701840f7148529e8e9ba240164edb908c3f12bb55519b1e"} Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.764959 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 11:01:11 crc kubenswrapper[4846]: I0320 11:01:11.764965 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1dd2c480593c5e8701840f7148529e8e9ba240164edb908c3f12bb55519b1e" Mar 20 11:01:12 crc kubenswrapper[4846]: I0320 11:01:12.541646 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7zwsq" Mar 20 11:01:13 crc kubenswrapper[4846]: I0320 11:01:13.045318 4846 ???:1] "http: TLS handshake error from 192.168.126.11:36490: no serving certificate available for the kubelet" Mar 20 11:01:16 crc kubenswrapper[4846]: I0320 11:01:16.382738 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-fkvwp" Mar 20 11:01:16 crc kubenswrapper[4846]: I0320 11:01:16.899973 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-d2pll_0a35745e-3b0d-44fa-a0d2-3f44b9f437ed/cluster-samples-operator/0.log" Mar 20 11:01:16 crc kubenswrapper[4846]: I0320 11:01:16.900457 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a35745e-3b0d-44fa-a0d2-3f44b9f437ed" containerID="f332f46c0adbb53c25bf3aa568df7224d3ddccc25f0cb812e8d33c17f2b2e22d" exitCode=2 Mar 20 11:01:16 crc kubenswrapper[4846]: I0320 11:01:16.900500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" event={"ID":"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed","Type":"ContainerDied","Data":"f332f46c0adbb53c25bf3aa568df7224d3ddccc25f0cb812e8d33c17f2b2e22d"} Mar 20 11:01:16 crc kubenswrapper[4846]: I0320 11:01:16.901282 4846 scope.go:117] "RemoveContainer" containerID="f332f46c0adbb53c25bf3aa568df7224d3ddccc25f0cb812e8d33c17f2b2e22d" Mar 20 11:01:17 crc kubenswrapper[4846]: I0320 11:01:17.060727 4846 patch_prober.go:28] interesting pod/console-f9d7485db-bwgwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 20 11:01:17 crc kubenswrapper[4846]: I0320 11:01:17.060800 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bwgwb" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 20 11:01:22 crc kubenswrapper[4846]: I0320 11:01:22.330631 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:22 crc kubenswrapper[4846]: I0320 11:01:22.334594 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:22 crc kubenswrapper[4846]: I0320 11:01:22.334935 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" containerID="cri-o://57961d873c7cc9b9911d43cdf0f3d121f5d6b5e8aca8069865b06e8d2234d07b" gracePeriod=30 Mar 20 11:01:22 crc kubenswrapper[4846]: I0320 11:01:22.335298 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerName="controller-manager" containerID="cri-o://76689621b1edcf8e9f0a435c9a8953aed400570db3e2667fd1f3f5a1c2459fdb" gracePeriod=30 Mar 20 11:01:26 crc kubenswrapper[4846]: I0320 11:01:26.075125 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:01:27 crc kubenswrapper[4846]: I0320 11:01:27.064313 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:01:27 crc kubenswrapper[4846]: I0320 11:01:27.069420 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:01:27 crc kubenswrapper[4846]: I0320 11:01:27.812410 4846 patch_prober.go:28] interesting pod/route-controller-manager-67776fb75d-ntnms container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Mar 20 11:01:27 crc kubenswrapper[4846]: I0320 11:01:27.812496 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Mar 20 11:01:28 crc kubenswrapper[4846]: I0320 11:01:28.171522 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 11:01:30 crc kubenswrapper[4846]: I0320 11:01:30.777287 4846 patch_prober.go:28] interesting pod/controller-manager-7dddc88b74-pdsbg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 11:01:30 crc kubenswrapper[4846]: I0320 11:01:30.777828 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 11:01:33 crc kubenswrapper[4846]: I0320 11:01:33.549536 4846 ???:1] "http: TLS handshake error from 192.168.126.11:35804: no serving certificate available for the kubelet" Mar 20 11:01:34 crc kubenswrapper[4846]: I0320 11:01:34.253771 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-7dddc88b74-pdsbg_a8858bf7-b571-4de3-9bc3-82e527e9b8b5/controller-manager/0.log" Mar 20 11:01:34 crc kubenswrapper[4846]: I0320 11:01:34.253868 4846 generic.go:334] "Generic (PLEG): container finished" podID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerID="76689621b1edcf8e9f0a435c9a8953aed400570db3e2667fd1f3f5a1c2459fdb" exitCode=-1 Mar 20 11:01:34 crc kubenswrapper[4846]: I0320 11:01:34.253971 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" event={"ID":"a8858bf7-b571-4de3-9bc3-82e527e9b8b5","Type":"ContainerDied","Data":"76689621b1edcf8e9f0a435c9a8953aed400570db3e2667fd1f3f5a1c2459fdb"} Mar 20 11:01:34 crc kubenswrapper[4846]: I0320 11:01:34.765362 4846 generic.go:334] "Generic (PLEG): container finished" podID="5806af56-cc91-46ec-9255-034a44f08001" containerID="57961d873c7cc9b9911d43cdf0f3d121f5d6b5e8aca8069865b06e8d2234d07b" exitCode=0 Mar 20 11:01:34 crc kubenswrapper[4846]: I0320 11:01:34.765475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" event={"ID":"5806af56-cc91-46ec-9255-034a44f08001","Type":"ContainerDied","Data":"57961d873c7cc9b9911d43cdf0f3d121f5d6b5e8aca8069865b06e8d2234d07b"} Mar 20 11:01:36 crc kubenswrapper[4846]: I0320 11:01:36.827645 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-kwbqc" Mar 20 11:01:38 crc kubenswrapper[4846]: I0320 11:01:38.812785 4846 patch_prober.go:28] interesting pod/route-controller-manager-67776fb75d-ntnms container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 11:01:38 crc kubenswrapper[4846]: I0320 11:01:38.813223 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.422230 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 11:01:39 crc kubenswrapper[4846]: E0320 11:01:39.423020 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32b5559-d9ec-4b43-ba99-c80678cf5e41" containerName="pruner" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.423038 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32b5559-d9ec-4b43-ba99-c80678cf5e41" containerName="pruner" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.423432 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32b5559-d9ec-4b43-ba99-c80678cf5e41" containerName="pruner" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.424378 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.428109 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.433432 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.435215 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.442101 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.442181 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.543824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.543929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.544079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.565627 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.677216 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.677323 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.677597 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.678275 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.720466 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:39 crc kubenswrapper[4846]: E0320 11:01:39.720764 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerName="controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.720781 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerName="controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: E0320 11:01:39.722473 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.722507 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.724547 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" containerName="controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.724580 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5806af56-cc91-46ec-9255-034a44f08001" containerName="route-controller-manager" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.725193 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.729934 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.745780 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config\") pod \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.745880 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config\") pod \"5806af56-cc91-46ec-9255-034a44f08001\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.745986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles\") pod \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.747709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config" (OuterVolumeSpecName: "config") pod "5806af56-cc91-46ec-9255-034a44f08001" (UID: "5806af56-cc91-46ec-9255-034a44f08001"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.747730 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a8858bf7-b571-4de3-9bc3-82e527e9b8b5" (UID: "a8858bf7-b571-4de3-9bc3-82e527e9b8b5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748270 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config" (OuterVolumeSpecName: "config") pod "a8858bf7-b571-4de3-9bc3-82e527e9b8b5" (UID: "a8858bf7-b571-4de3-9bc3-82e527e9b8b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqdgw\" (UniqueName: \"kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw\") pod \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca\") pod \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748602 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5qms\" (UniqueName: \"kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms\") pod \"5806af56-cc91-46ec-9255-034a44f08001\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748662 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert\") pod \"5806af56-cc91-46ec-9255-034a44f08001\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748701 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca\") pod \"5806af56-cc91-46ec-9255-034a44f08001\" (UID: \"5806af56-cc91-46ec-9255-034a44f08001\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.748728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert\") pod \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\" (UID: \"a8858bf7-b571-4de3-9bc3-82e527e9b8b5\") " Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749131 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749177 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca" (OuterVolumeSpecName: "client-ca") pod "a8858bf7-b571-4de3-9bc3-82e527e9b8b5" (UID: "a8858bf7-b571-4de3-9bc3-82e527e9b8b5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5cnd\" (UniqueName: \"kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca" (OuterVolumeSpecName: "client-ca") pod "5806af56-cc91-46ec-9255-034a44f08001" (UID: "5806af56-cc91-46ec-9255-034a44f08001"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749615 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749659 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749682 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.749700 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.753614 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.754454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a8858bf7-b571-4de3-9bc3-82e527e9b8b5" (UID: "a8858bf7-b571-4de3-9bc3-82e527e9b8b5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.756079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5806af56-cc91-46ec-9255-034a44f08001" (UID: "5806af56-cc91-46ec-9255-034a44f08001"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.757096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw" (OuterVolumeSpecName: "kube-api-access-rqdgw") pod "a8858bf7-b571-4de3-9bc3-82e527e9b8b5" (UID: "a8858bf7-b571-4de3-9bc3-82e527e9b8b5"). InnerVolumeSpecName "kube-api-access-rqdgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.760323 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms" (OuterVolumeSpecName: "kube-api-access-q5qms") pod "5806af56-cc91-46ec-9255-034a44f08001" (UID: "5806af56-cc91-46ec-9255-034a44f08001"). InnerVolumeSpecName "kube-api-access-q5qms". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.804307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" event={"ID":"a8858bf7-b571-4de3-9bc3-82e527e9b8b5","Type":"ContainerDied","Data":"e361f54d025ed569e5befd62b7e405f71f18ad826fdf3a6b5c67989e60020ba4"} Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.804452 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7dddc88b74-pdsbg" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.804472 4846 scope.go:117] "RemoveContainer" containerID="76689621b1edcf8e9f0a435c9a8953aed400570db3e2667fd1f3f5a1c2459fdb" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.807599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" event={"ID":"5806af56-cc91-46ec-9255-034a44f08001","Type":"ContainerDied","Data":"53ad959f044db20eef968c12b3af3f0f57966f3854fd792f9cd12f59c1a9f9cb"} Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.807728 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.840341 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.845822 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67776fb75d-ntnms"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5cnd\" (UniqueName: \"kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855425 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855540 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqdgw\" (UniqueName: \"kubernetes.io/projected/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-kube-api-access-rqdgw\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855560 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5qms\" (UniqueName: \"kubernetes.io/projected/5806af56-cc91-46ec-9255-034a44f08001-kube-api-access-q5qms\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855574 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5806af56-cc91-46ec-9255-034a44f08001-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855588 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5806af56-cc91-46ec-9255-034a44f08001-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.855650 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8858bf7-b571-4de3-9bc3-82e527e9b8b5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.856880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.857122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.858271 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.859137 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.861449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.862846 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7dddc88b74-pdsbg"] Mar 20 11:01:39 crc kubenswrapper[4846]: I0320 11:01:39.876515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5cnd\" (UniqueName: \"kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd\") pod \"controller-manager-c58948574-zqxsc\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:40 crc kubenswrapper[4846]: I0320 11:01:40.050649 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:41 crc kubenswrapper[4846]: I0320 11:01:41.333974 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5806af56-cc91-46ec-9255-034a44f08001" path="/var/lib/kubelet/pods/5806af56-cc91-46ec-9255-034a44f08001/volumes" Mar 20 11:01:41 crc kubenswrapper[4846]: I0320 11:01:41.335603 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8858bf7-b571-4de3-9bc3-82e527e9b8b5" path="/var/lib/kubelet/pods/a8858bf7-b571-4de3-9bc3-82e527e9b8b5/volumes" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.361264 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.494874 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.496407 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.499828 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.500105 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.500214 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.500421 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.500511 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.500742 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.508589 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.604937 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.605049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.605092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.605123 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjhjc\" (UniqueName: \"kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.706167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.706263 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.706289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.706308 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjhjc\" (UniqueName: \"kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.709371 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.710040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.717272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.723779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjhjc\" (UniqueName: \"kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc\") pod \"route-controller-manager-84b545b858-5mq8f\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:42 crc kubenswrapper[4846]: I0320 11:01:42.817585 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.339139 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.340050 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xjvlg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kh2d5_openshift-marketplace(de1fb688-1d64-4289-9d68-faf88d479fa1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.341257 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kh2d5" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.344840 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.344977 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r9pfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nhrbq_openshift-marketplace(f404899d-1c18-424f-ab1c-2c7c96c2e368): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:44 crc kubenswrapper[4846]: E0320 11:01:44.346466 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nhrbq" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.410221 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.411452 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.414318 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.535793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.535926 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.535946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.638042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.638095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.638151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.638232 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.638230 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.660767 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access\") pod \"installer-9-crc\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:44 crc kubenswrapper[4846]: I0320 11:01:44.739606 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:01:46 crc kubenswrapper[4846]: E0320 11:01:46.310981 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kh2d5" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" Mar 20 11:01:46 crc kubenswrapper[4846]: E0320 11:01:46.402728 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 11:01:46 crc kubenswrapper[4846]: E0320 11:01:46.403038 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q7dfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4z7gv_openshift-marketplace(6ba6d606-56e4-4285-887a-4b86f8198046): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:46 crc kubenswrapper[4846]: E0320 11:01:46.404254 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4z7gv" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" Mar 20 11:01:48 crc kubenswrapper[4846]: E0320 11:01:48.046588 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4z7gv" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" Mar 20 11:01:48 crc kubenswrapper[4846]: E0320 11:01:48.118043 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 11:01:48 crc kubenswrapper[4846]: E0320 11:01:48.118281 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgwxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lh8ff_openshift-marketplace(408ee9c3-1b20-4002-850f-328b27b972d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:48 crc kubenswrapper[4846]: E0320 11:01:48.119500 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lh8ff" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.474588 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lh8ff" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.542663 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.542873 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p45cb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6vwj9_openshift-marketplace(bb0fabe2-dd2f-451f-8f04-68feddb0fdd7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.545068 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6vwj9" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.573657 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.573885 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4s4jw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-56jhc_openshift-marketplace(2b17dfdf-1984-4cde-958e-021958ad8e0c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.575166 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-56jhc" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.600918 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.601249 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vj6gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tt9fr_openshift-marketplace(d6291c67-3539-45a8-bbb5-312be6bebcd4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:49 crc kubenswrapper[4846]: E0320 11:01:49.602385 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tt9fr" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.454450 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-56jhc" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.454974 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tt9fr" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.455220 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6vwj9" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.483202 4846 scope.go:117] "RemoveContainer" containerID="57961d873c7cc9b9911d43cdf0f3d121f5d6b5e8aca8069865b06e8d2234d07b" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.493736 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.494014 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 11:01:50 crc kubenswrapper[4846]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 20 11:01:50 crc kubenswrapper[4846]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9whd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29566740-46wjt_openshift-infra(8dd31859-c683-4f1a-b167-3f20a96b6f9c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 20 11:01:50 crc kubenswrapper[4846]: > logger="UnhandledError" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.495489 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29566740-46wjt" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.545711 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.546406 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5mq5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6qggw_openshift-marketplace(131db4bd-a795-401c-a9e7-b66f4512dc32): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.548502 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6qggw" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.877562 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:01:50 crc kubenswrapper[4846]: W0320 11:01:50.882036 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod950ceea7_d8ec_4a01_a48f_977003a9216d.slice/crio-dbf759db2b3aa555b4a8ae51ba785abfd415d9221c4d7e3fb9c16bebe8bec334 WatchSource:0}: Error finding container dbf759db2b3aa555b4a8ae51ba785abfd415d9221c4d7e3fb9c16bebe8bec334: Status 404 returned error can't find the container with id dbf759db2b3aa555b4a8ae51ba785abfd415d9221c4d7e3fb9c16bebe8bec334 Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.887406 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-d2pll_0a35745e-3b0d-44fa-a0d2-3f44b9f437ed/cluster-samples-operator/0.log" Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.887494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d2pll" event={"ID":"0a35745e-3b0d-44fa-a0d2-3f44b9f437ed","Type":"ContainerStarted","Data":"eb493f5730a3633caeec7df364d3153495c40e285325af5b9b58e64966aa7b02"} Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.904216 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29566740-46wjt" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" Mar 20 11:01:50 crc kubenswrapper[4846]: E0320 11:01:50.905572 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6qggw" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.937486 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.952812 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 11:01:50 crc kubenswrapper[4846]: I0320 11:01:50.963208 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:51 crc kubenswrapper[4846]: W0320 11:01:51.003134 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod635a1db4_bbe5_41bb_9591_e9b763abe8ce.slice/crio-b24bafd7e75595b6e6d4a7d580ff5635a9e59d4b401bf13beb22eb2d6c022d0c WatchSource:0}: Error finding container b24bafd7e75595b6e6d4a7d580ff5635a9e59d4b401bf13beb22eb2d6c022d0c: Status 404 returned error can't find the container with id b24bafd7e75595b6e6d4a7d580ff5635a9e59d4b401bf13beb22eb2d6c022d0c Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.904999 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3aa37f4d-5b65-45f4-ae84-a1f563508ff6","Type":"ContainerStarted","Data":"b1eccb669b422d66113b2a4ed1138d23c4122eabf9098002c4ae503e15b25e0c"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.905992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3aa37f4d-5b65-45f4-ae84-a1f563508ff6","Type":"ContainerStarted","Data":"88d2027a148d51d187584d9034c1ddb74e175f33e84b6c5e22802a0f7833b910"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.906477 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" event={"ID":"950ceea7-d8ec-4a01-a48f-977003a9216d","Type":"ContainerStarted","Data":"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.906532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" event={"ID":"950ceea7-d8ec-4a01-a48f-977003a9216d","Type":"ContainerStarted","Data":"dbf759db2b3aa555b4a8ae51ba785abfd415d9221c4d7e3fb9c16bebe8bec334"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.906989 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.908315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c0efeb5d-46d6-4f9b-b563-70aa36f58062","Type":"ContainerStarted","Data":"0865ef647eaac8cd9aa5f596f1eb739bf92ee813dddd26397366937d5789bf12"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.908344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c0efeb5d-46d6-4f9b-b563-70aa36f58062","Type":"ContainerStarted","Data":"02934d31308f89cfb103c32ddab2a2c2dce2df2b15b2784e1e525468d2d85c61"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.912673 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" event={"ID":"635a1db4-bbe5-41bb-9591-e9b763abe8ce","Type":"ContainerStarted","Data":"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.912827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" event={"ID":"635a1db4-bbe5-41bb-9591-e9b763abe8ce","Type":"ContainerStarted","Data":"b24bafd7e75595b6e6d4a7d580ff5635a9e59d4b401bf13beb22eb2d6c022d0c"} Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.913027 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" podUID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" containerName="controller-manager" containerID="cri-o://53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98" gracePeriod=30 Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.913625 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.913698 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.923982 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.930253 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.930227043 podStartE2EDuration="7.930227043s" podCreationTimestamp="2026-03-20 11:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:51.927794577 +0000 UTC m=+283.294202813" watchObservedRunningTime="2026-03-20 11:01:51.930227043 +0000 UTC m=+283.296635279" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.949192 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=12.949165835 podStartE2EDuration="12.949165835s" podCreationTimestamp="2026-03-20 11:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:51.945874726 +0000 UTC m=+283.312282962" watchObservedRunningTime="2026-03-20 11:01:51.949165835 +0000 UTC m=+283.315574071" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.973717 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" podStartSLOduration=29.973694278 podStartE2EDuration="29.973694278s" podCreationTimestamp="2026-03-20 11:01:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:51.973433891 +0000 UTC m=+283.339842127" watchObservedRunningTime="2026-03-20 11:01:51.973694278 +0000 UTC m=+283.340102584" Mar 20 11:01:51 crc kubenswrapper[4846]: I0320 11:01:51.996691 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" podStartSLOduration=9.996672909 podStartE2EDuration="9.996672909s" podCreationTimestamp="2026-03-20 11:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:51.993582805 +0000 UTC m=+283.359991071" watchObservedRunningTime="2026-03-20 11:01:51.996672909 +0000 UTC m=+283.363081145" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.280208 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.305246 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:01:52 crc kubenswrapper[4846]: E0320 11:01:52.305548 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" containerName="controller-manager" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.305562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" containerName="controller-manager" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.305674 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" containerName="controller-manager" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.306730 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.330118 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.368437 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5cnd\" (UniqueName: \"kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd\") pod \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.368772 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca\") pod \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.368998 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.369112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.369199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.369289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.369356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4fdc\" (UniqueName: \"kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.369741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca" (OuterVolumeSpecName: "client-ca") pod "635a1db4-bbe5-41bb-9591-e9b763abe8ce" (UID: "635a1db4-bbe5-41bb-9591-e9b763abe8ce"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.374741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd" (OuterVolumeSpecName: "kube-api-access-v5cnd") pod "635a1db4-bbe5-41bb-9591-e9b763abe8ce" (UID: "635a1db4-bbe5-41bb-9591-e9b763abe8ce"). InnerVolumeSpecName "kube-api-access-v5cnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470267 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config\") pod \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470341 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert\") pod \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470406 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles\") pod \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\" (UID: \"635a1db4-bbe5-41bb-9591-e9b763abe8ce\") " Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470910 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4fdc\" (UniqueName: \"kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.470988 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471011 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config" (OuterVolumeSpecName: "config") pod "635a1db4-bbe5-41bb-9591-e9b763abe8ce" (UID: "635a1db4-bbe5-41bb-9591-e9b763abe8ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471134 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5cnd\" (UniqueName: \"kubernetes.io/projected/635a1db4-bbe5-41bb-9591-e9b763abe8ce-kube-api-access-v5cnd\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471151 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471165 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.471442 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "635a1db4-bbe5-41bb-9591-e9b763abe8ce" (UID: "635a1db4-bbe5-41bb-9591-e9b763abe8ce"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.472582 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.472652 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.473827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.475193 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "635a1db4-bbe5-41bb-9591-e9b763abe8ce" (UID: "635a1db4-bbe5-41bb-9591-e9b763abe8ce"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.476569 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.495668 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4fdc\" (UniqueName: \"kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc\") pod \"controller-manager-5c98cddd45-bwsbz\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.571705 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/635a1db4-bbe5-41bb-9591-e9b763abe8ce-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.571746 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/635a1db4-bbe5-41bb-9591-e9b763abe8ce-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.630278 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.826312 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.917942 4846 generic.go:334] "Generic (PLEG): container finished" podID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" containerID="53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98" exitCode=0 Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.918006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" event={"ID":"635a1db4-bbe5-41bb-9591-e9b763abe8ce","Type":"ContainerDied","Data":"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98"} Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.918037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" event={"ID":"635a1db4-bbe5-41bb-9591-e9b763abe8ce","Type":"ContainerDied","Data":"b24bafd7e75595b6e6d4a7d580ff5635a9e59d4b401bf13beb22eb2d6c022d0c"} Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.918058 4846 scope.go:117] "RemoveContainer" containerID="53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.918180 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c58948574-zqxsc" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.934483 4846 generic.go:334] "Generic (PLEG): container finished" podID="c0efeb5d-46d6-4f9b-b563-70aa36f58062" containerID="0865ef647eaac8cd9aa5f596f1eb739bf92ee813dddd26397366937d5789bf12" exitCode=0 Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.934643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c0efeb5d-46d6-4f9b-b563-70aa36f58062","Type":"ContainerDied","Data":"0865ef647eaac8cd9aa5f596f1eb739bf92ee813dddd26397366937d5789bf12"} Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.941619 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" event={"ID":"044abdf3-a57e-4b57-a774-009fb7b1e5d2","Type":"ContainerStarted","Data":"eb1e9c42c1e7f93134f377a8bb24d00902391cae20050def8091c1ae99e4bfe6"} Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.957337 4846 scope.go:117] "RemoveContainer" containerID="53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98" Mar 20 11:01:52 crc kubenswrapper[4846]: E0320 11:01:52.958247 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98\": container with ID starting with 53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98 not found: ID does not exist" containerID="53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.958300 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98"} err="failed to get container status \"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98\": rpc error: code = NotFound desc = could not find container \"53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98\": container with ID starting with 53d94fb236ef08ab53bc61c5c578520e14168104c1e3db641c18f229d0eb4b98 not found: ID does not exist" Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.971448 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:52 crc kubenswrapper[4846]: I0320 11:01:52.975522 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-c58948574-zqxsc"] Mar 20 11:01:53 crc kubenswrapper[4846]: I0320 11:01:53.330703 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635a1db4-bbe5-41bb-9591-e9b763abe8ce" path="/var/lib/kubelet/pods/635a1db4-bbe5-41bb-9591-e9b763abe8ce/volumes" Mar 20 11:01:53 crc kubenswrapper[4846]: I0320 11:01:53.954145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" event={"ID":"044abdf3-a57e-4b57-a774-009fb7b1e5d2","Type":"ContainerStarted","Data":"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af"} Mar 20 11:01:53 crc kubenswrapper[4846]: I0320 11:01:53.954778 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:53 crc kubenswrapper[4846]: I0320 11:01:53.962584 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:01:53 crc kubenswrapper[4846]: I0320 11:01:53.995191 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" podStartSLOduration=11.995159374 podStartE2EDuration="11.995159374s" podCreationTimestamp="2026-03-20 11:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:01:53.97539215 +0000 UTC m=+285.341800386" watchObservedRunningTime="2026-03-20 11:01:53.995159374 +0000 UTC m=+285.361567610" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.226017 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.398739 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access\") pod \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.398918 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir\") pod \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\" (UID: \"c0efeb5d-46d6-4f9b-b563-70aa36f58062\") " Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.399110 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c0efeb5d-46d6-4f9b-b563-70aa36f58062" (UID: "c0efeb5d-46d6-4f9b-b563-70aa36f58062"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.405217 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c0efeb5d-46d6-4f9b-b563-70aa36f58062" (UID: "c0efeb5d-46d6-4f9b-b563-70aa36f58062"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.500399 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.500440 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0efeb5d-46d6-4f9b-b563-70aa36f58062-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.963121 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"c0efeb5d-46d6-4f9b-b563-70aa36f58062","Type":"ContainerDied","Data":"02934d31308f89cfb103c32ddab2a2c2dce2df2b15b2784e1e525468d2d85c61"} Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.963176 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02934d31308f89cfb103c32ddab2a2c2dce2df2b15b2784e1e525468d2d85c61" Mar 20 11:01:54 crc kubenswrapper[4846]: I0320 11:01:54.963229 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 11:01:57 crc kubenswrapper[4846]: I0320 11:01:57.982494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerStarted","Data":"11c3f05d48489567e31bbe3dd4ab70142418e8d41c983b4986c99498f7a8fb4e"} Mar 20 11:01:58 crc kubenswrapper[4846]: I0320 11:01:58.992555 4846 generic.go:334] "Generic (PLEG): container finished" podID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerID="11c3f05d48489567e31bbe3dd4ab70142418e8d41c983b4986c99498f7a8fb4e" exitCode=0 Mar 20 11:01:58 crc kubenswrapper[4846]: I0320 11:01:58.992667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerDied","Data":"11c3f05d48489567e31bbe3dd4ab70142418e8d41c983b4986c99498f7a8fb4e"} Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.007598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerStarted","Data":"3d08ab3d8c6a5543e359243a4c72f68ce419280ab30fd73f67f7e8bf9403c88f"} Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.010563 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerStarted","Data":"95fbc174cd3771e5ed2f271d26205b0ab38e476459b88aed13572bf7a26c2cc3"} Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.060495 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nhrbq" podStartSLOduration=4.407317297 podStartE2EDuration="53.060466975s" podCreationTimestamp="2026-03-20 11:01:07 +0000 UTC" firstStartedPulling="2026-03-20 11:01:10.726982279 +0000 UTC m=+242.093390515" lastFinishedPulling="2026-03-20 11:01:59.380131947 +0000 UTC m=+290.746540193" observedRunningTime="2026-03-20 11:02:00.058753717 +0000 UTC m=+291.425161963" watchObservedRunningTime="2026-03-20 11:02:00.060466975 +0000 UTC m=+291.426875211" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.141912 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566742-wqk8p"] Mar 20 11:02:00 crc kubenswrapper[4846]: E0320 11:02:00.142490 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0efeb5d-46d6-4f9b-b563-70aa36f58062" containerName="pruner" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.142590 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0efeb5d-46d6-4f9b-b563-70aa36f58062" containerName="pruner" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.142785 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0efeb5d-46d6-4f9b-b563-70aa36f58062" containerName="pruner" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.143380 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.148716 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.152973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566742-wqk8p"] Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.187207 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhf86\" (UniqueName: \"kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86\") pod \"auto-csr-approver-29566742-wqk8p\" (UID: \"fe1698aa-b5c6-4480-a886-c89481d08c59\") " pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.288826 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhf86\" (UniqueName: \"kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86\") pod \"auto-csr-approver-29566742-wqk8p\" (UID: \"fe1698aa-b5c6-4480-a886-c89481d08c59\") " pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.314616 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhf86\" (UniqueName: \"kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86\") pod \"auto-csr-approver-29566742-wqk8p\" (UID: \"fe1698aa-b5c6-4480-a886-c89481d08c59\") " pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.461336 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:00 crc kubenswrapper[4846]: I0320 11:02:00.894721 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566742-wqk8p"] Mar 20 11:02:00 crc kubenswrapper[4846]: W0320 11:02:00.912197 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe1698aa_b5c6_4480_a886_c89481d08c59.slice/crio-ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602 WatchSource:0}: Error finding container ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602: Status 404 returned error can't find the container with id ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602 Mar 20 11:02:01 crc kubenswrapper[4846]: I0320 11:02:01.020120 4846 generic.go:334] "Generic (PLEG): container finished" podID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerID="3d08ab3d8c6a5543e359243a4c72f68ce419280ab30fd73f67f7e8bf9403c88f" exitCode=0 Mar 20 11:02:01 crc kubenswrapper[4846]: I0320 11:02:01.020221 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerDied","Data":"3d08ab3d8c6a5543e359243a4c72f68ce419280ab30fd73f67f7e8bf9403c88f"} Mar 20 11:02:01 crc kubenswrapper[4846]: I0320 11:02:01.021595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" event={"ID":"fe1698aa-b5c6-4480-a886-c89481d08c59","Type":"ContainerStarted","Data":"ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602"} Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.032384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerStarted","Data":"7694dc1764c7328d2c31062299e1d6cd0fe863b3ba7083df719e4d4d21900533"} Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.328978 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kh2d5" podStartSLOduration=3.533693898 podStartE2EDuration="56.328957896s" podCreationTimestamp="2026-03-20 11:01:06 +0000 UTC" firstStartedPulling="2026-03-20 11:01:08.636593691 +0000 UTC m=+240.003001927" lastFinishedPulling="2026-03-20 11:02:01.431857689 +0000 UTC m=+292.798265925" observedRunningTime="2026-03-20 11:02:02.058762424 +0000 UTC m=+293.425170670" watchObservedRunningTime="2026-03-20 11:02:02.328957896 +0000 UTC m=+293.695366132" Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.332378 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.332648 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerName="controller-manager" containerID="cri-o://c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af" gracePeriod=30 Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.373806 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.374101 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerName="route-controller-manager" containerID="cri-o://7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4" gracePeriod=30 Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.918028 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.928400 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjhjc\" (UniqueName: \"kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc\") pod \"950ceea7-d8ec-4a01-a48f-977003a9216d\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.928455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca\") pod \"950ceea7-d8ec-4a01-a48f-977003a9216d\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.928513 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert\") pod \"950ceea7-d8ec-4a01-a48f-977003a9216d\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.928560 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config\") pod \"950ceea7-d8ec-4a01-a48f-977003a9216d\" (UID: \"950ceea7-d8ec-4a01-a48f-977003a9216d\") " Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.929823 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config" (OuterVolumeSpecName: "config") pod "950ceea7-d8ec-4a01-a48f-977003a9216d" (UID: "950ceea7-d8ec-4a01-a48f-977003a9216d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.930185 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca" (OuterVolumeSpecName: "client-ca") pod "950ceea7-d8ec-4a01-a48f-977003a9216d" (UID: "950ceea7-d8ec-4a01-a48f-977003a9216d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.937285 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc" (OuterVolumeSpecName: "kube-api-access-wjhjc") pod "950ceea7-d8ec-4a01-a48f-977003a9216d" (UID: "950ceea7-d8ec-4a01-a48f-977003a9216d"). InnerVolumeSpecName "kube-api-access-wjhjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:02 crc kubenswrapper[4846]: I0320 11:02:02.939214 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "950ceea7-d8ec-4a01-a48f-977003a9216d" (UID: "950ceea7-d8ec-4a01-a48f-977003a9216d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.025742 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.029584 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles\") pod \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.029666 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca\") pod \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.029762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config\") pod \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.029854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert\") pod \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.029953 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4fdc\" (UniqueName: \"kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc\") pod \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\" (UID: \"044abdf3-a57e-4b57-a774-009fb7b1e5d2\") " Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.030274 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.030294 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjhjc\" (UniqueName: \"kubernetes.io/projected/950ceea7-d8ec-4a01-a48f-977003a9216d-kube-api-access-wjhjc\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.030309 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/950ceea7-d8ec-4a01-a48f-977003a9216d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.030319 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/950ceea7-d8ec-4a01-a48f-977003a9216d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.031207 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "044abdf3-a57e-4b57-a774-009fb7b1e5d2" (UID: "044abdf3-a57e-4b57-a774-009fb7b1e5d2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.031240 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca" (OuterVolumeSpecName: "client-ca") pod "044abdf3-a57e-4b57-a774-009fb7b1e5d2" (UID: "044abdf3-a57e-4b57-a774-009fb7b1e5d2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.031451 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config" (OuterVolumeSpecName: "config") pod "044abdf3-a57e-4b57-a774-009fb7b1e5d2" (UID: "044abdf3-a57e-4b57-a774-009fb7b1e5d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.036422 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "044abdf3-a57e-4b57-a774-009fb7b1e5d2" (UID: "044abdf3-a57e-4b57-a774-009fb7b1e5d2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.048603 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ba6d606-56e4-4285-887a-4b86f8198046" containerID="95dad7bc2edb7648d231390bb4815733ea47c81d9e138016e60003b804692b61" exitCode=0 Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.048678 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerDied","Data":"95dad7bc2edb7648d231390bb4815733ea47c81d9e138016e60003b804692b61"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.051719 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc" (OuterVolumeSpecName: "kube-api-access-t4fdc") pod "044abdf3-a57e-4b57-a774-009fb7b1e5d2" (UID: "044abdf3-a57e-4b57-a774-009fb7b1e5d2"). InnerVolumeSpecName "kube-api-access-t4fdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.058166 4846 generic.go:334] "Generic (PLEG): container finished" podID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerID="7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4" exitCode=0 Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.058253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" event={"ID":"950ceea7-d8ec-4a01-a48f-977003a9216d","Type":"ContainerDied","Data":"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.058269 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.058292 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" event={"ID":"950ceea7-d8ec-4a01-a48f-977003a9216d","Type":"ContainerDied","Data":"dbf759db2b3aa555b4a8ae51ba785abfd415d9221c4d7e3fb9c16bebe8bec334"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.058314 4846 scope.go:117] "RemoveContainer" containerID="7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.067816 4846 generic.go:334] "Generic (PLEG): container finished" podID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerID="c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af" exitCode=0 Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.068012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" event={"ID":"044abdf3-a57e-4b57-a774-009fb7b1e5d2","Type":"ContainerDied","Data":"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.068120 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" event={"ID":"044abdf3-a57e-4b57-a774-009fb7b1e5d2","Type":"ContainerDied","Data":"eb1e9c42c1e7f93134f377a8bb24d00902391cae20050def8091c1ae99e4bfe6"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.068326 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.073036 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" event={"ID":"fe1698aa-b5c6-4480-a886-c89481d08c59","Type":"ContainerStarted","Data":"a3eacf9b9ba175458f62f229ad9be0b1beb42084de63bb452faf0dbc5168f703"} Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.096266 4846 scope.go:117] "RemoveContainer" containerID="7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4" Mar 20 11:02:03 crc kubenswrapper[4846]: E0320 11:02:03.107639 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4\": container with ID starting with 7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4 not found: ID does not exist" containerID="7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.107694 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4"} err="failed to get container status \"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4\": rpc error: code = NotFound desc = could not find container \"7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4\": container with ID starting with 7c8a2d0e359526053a4b985cbb5a88f8af1d99f7b0342a9883f5c71535c446d4 not found: ID does not exist" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.107729 4846 scope.go:117] "RemoveContainer" containerID="c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.125269 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" podStartSLOduration=1.889055506 podStartE2EDuration="3.125246678s" podCreationTimestamp="2026-03-20 11:02:00 +0000 UTC" firstStartedPulling="2026-03-20 11:02:00.9175821 +0000 UTC m=+292.283990336" lastFinishedPulling="2026-03-20 11:02:02.153773252 +0000 UTC m=+293.520181508" observedRunningTime="2026-03-20 11:02:03.122541325 +0000 UTC m=+294.488949561" watchObservedRunningTime="2026-03-20 11:02:03.125246678 +0000 UTC m=+294.491654904" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.147230 4846 scope.go:117] "RemoveContainer" containerID="c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af" Mar 20 11:02:03 crc kubenswrapper[4846]: E0320 11:02:03.152823 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af\": container with ID starting with c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af not found: ID does not exist" containerID="c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.153042 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af"} err="failed to get container status \"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af\": rpc error: code = NotFound desc = could not find container \"c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af\": container with ID starting with c34a43f3c3c85644aa9f634b9e682f7ac3a789601ca773bafde880feb0bf94af not found: ID does not exist" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.159731 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/044abdf3-a57e-4b57-a774-009fb7b1e5d2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.162834 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4fdc\" (UniqueName: \"kubernetes.io/projected/044abdf3-a57e-4b57-a774-009fb7b1e5d2-kube-api-access-t4fdc\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.162863 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.162885 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.162913 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/044abdf3-a57e-4b57-a774-009fb7b1e5d2-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.168530 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.175641 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c98cddd45-bwsbz"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.179759 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.182797 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.289857 4846 csr.go:261] certificate signing request csr-j85jv is approved, waiting to be issued Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.300205 4846 csr.go:257] certificate signing request csr-j85jv is issued Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.332032 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" path="/var/lib/kubelet/pods/044abdf3-a57e-4b57-a774-009fb7b1e5d2/volumes" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.332659 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" path="/var/lib/kubelet/pods/950ceea7-d8ec-4a01-a48f-977003a9216d/volumes" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.504720 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:03 crc kubenswrapper[4846]: E0320 11:02:03.505271 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerName="route-controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.505295 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerName="route-controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: E0320 11:02:03.505318 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerName="controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.505327 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerName="controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.505478 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerName="controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.505506 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerName="route-controller-manager" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.506243 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.507417 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.508548 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.510744 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.511217 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.511347 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.511347 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.511447 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.516359 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.516523 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.516737 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.516826 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.517276 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.517481 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.518127 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.525384 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.527177 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.532996 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568247 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568366 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568475 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62754\" (UniqueName: \"kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwg4f\" (UniqueName: \"kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.568848 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.631544 4846 patch_prober.go:28] interesting pod/controller-manager-5c98cddd45-bwsbz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.631680 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5c98cddd45-bwsbz" podUID="044abdf3-a57e-4b57-a774-009fb7b1e5d2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.670944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671049 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671083 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62754\" (UniqueName: \"kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671136 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwg4f\" (UniqueName: \"kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671178 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.671238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.672325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.672387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.673647 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.677187 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.677756 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.680879 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.687597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.698351 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62754\" (UniqueName: \"kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754\") pod \"controller-manager-679bd75db9-w7fqd\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.700326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwg4f\" (UniqueName: \"kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f\") pod \"route-controller-manager-5bcf5d65c9-sxx5z\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.819085 4846 patch_prober.go:28] interesting pod/route-controller-manager-84b545b858-5mq8f container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.819209 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-84b545b858-5mq8f" podUID="950ceea7-d8ec-4a01-a48f-977003a9216d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.830336 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:03 crc kubenswrapper[4846]: I0320 11:02:03.841029 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.111367 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:04 crc kubenswrapper[4846]: W0320 11:02:04.171723 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod875040a8_922a_4658_b3ec_205a1341d541.slice/crio-aa4aeb22c6c968d4a51d222905a69295c07e39a189f1b0b4e07aa2b322f2d8d4 WatchSource:0}: Error finding container aa4aeb22c6c968d4a51d222905a69295c07e39a189f1b0b4e07aa2b322f2d8d4: Status 404 returned error can't find the container with id aa4aeb22c6c968d4a51d222905a69295c07e39a189f1b0b4e07aa2b322f2d8d4 Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.175675 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerStarted","Data":"61386f7d4a6aef54983984e41c8e91c231429f0f56c6fee7bedfd72caa6f0395"} Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.201173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerStarted","Data":"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe"} Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.207940 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe1698aa-b5c6-4480-a886-c89481d08c59" containerID="a3eacf9b9ba175458f62f229ad9be0b1beb42084de63bb452faf0dbc5168f703" exitCode=0 Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.207992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" event={"ID":"fe1698aa-b5c6-4480-a886-c89481d08c59","Type":"ContainerDied","Data":"a3eacf9b9ba175458f62f229ad9be0b1beb42084de63bb452faf0dbc5168f703"} Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.234589 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4z7gv" podStartSLOduration=2.830925367 podStartE2EDuration="1m0.2345637s" podCreationTimestamp="2026-03-20 11:01:04 +0000 UTC" firstStartedPulling="2026-03-20 11:01:06.387634366 +0000 UTC m=+237.754042602" lastFinishedPulling="2026-03-20 11:02:03.791272699 +0000 UTC m=+295.157680935" observedRunningTime="2026-03-20 11:02:04.204819977 +0000 UTC m=+295.571228213" watchObservedRunningTime="2026-03-20 11:02:04.2345637 +0000 UTC m=+295.600971936" Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.240101 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.301449 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-25 15:44:34.612025978 +0000 UTC Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.301509 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6724h42m30.310520738s for next certificate rotation Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.738443 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:04 crc kubenswrapper[4846]: I0320 11:02:04.738928 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.215031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566740-46wjt" event={"ID":"8dd31859-c683-4f1a-b167-3f20a96b6f9c","Type":"ContainerStarted","Data":"4bfa365c6112a2b52c8b380c45beec1408fd692036f0d3fd40c624ec3b03310e"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.217280 4846 generic.go:334] "Generic (PLEG): container finished" podID="408ee9c3-1b20-4002-850f-328b27b972d0" containerID="971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe" exitCode=0 Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.217340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerDied","Data":"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.219755 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" event={"ID":"875040a8-922a-4658-b3ec-205a1341d541","Type":"ContainerStarted","Data":"2022d6a1102d6c110d784b753fe69b40d8116d69e5a4e6e52e5206797c3251fd"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.219789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" event={"ID":"875040a8-922a-4658-b3ec-205a1341d541","Type":"ContainerStarted","Data":"aa4aeb22c6c968d4a51d222905a69295c07e39a189f1b0b4e07aa2b322f2d8d4"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.221218 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.223966 4846 generic.go:334] "Generic (PLEG): container finished" podID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerID="810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8" exitCode=0 Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.224109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerDied","Data":"810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.225795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" event={"ID":"bc42705d-a520-4db2-994b-c626d1ed2e34","Type":"ContainerStarted","Data":"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.225822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" event={"ID":"bc42705d-a520-4db2-994b-c626d1ed2e34","Type":"ContainerStarted","Data":"87a1910f5cb4e0866fa602a00d6175d5460033f002f9c868284d2454498efc5c"} Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.226413 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.229109 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.234098 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.248348 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" podStartSLOduration=3.24832885 podStartE2EDuration="3.24832885s" podCreationTimestamp="2026-03-20 11:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:02:05.246385047 +0000 UTC m=+296.612793283" watchObservedRunningTime="2026-03-20 11:02:05.24832885 +0000 UTC m=+296.614737086" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.294501 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" podStartSLOduration=3.294480307 podStartE2EDuration="3.294480307s" podCreationTimestamp="2026-03-20 11:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:02:05.290851239 +0000 UTC m=+296.657259485" watchObservedRunningTime="2026-03-20 11:02:05.294480307 +0000 UTC m=+296.660888533" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.302037 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-20 15:09:29.643377877 +0000 UTC Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.302079 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6604h7m24.341301075s for next certificate rotation Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.688694 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.729559 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhf86\" (UniqueName: \"kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86\") pod \"fe1698aa-b5c6-4480-a886-c89481d08c59\" (UID: \"fe1698aa-b5c6-4480-a886-c89481d08c59\") " Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.737342 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86" (OuterVolumeSpecName: "kube-api-access-mhf86") pod "fe1698aa-b5c6-4480-a886-c89481d08c59" (UID: "fe1698aa-b5c6-4480-a886-c89481d08c59"). InnerVolumeSpecName "kube-api-access-mhf86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:05 crc kubenswrapper[4846]: I0320 11:02:05.831675 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhf86\" (UniqueName: \"kubernetes.io/projected/fe1698aa-b5c6-4480-a886-c89481d08c59-kube-api-access-mhf86\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.232918 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerStarted","Data":"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.234313 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerID="39f3bd4df3b33bcdda9be40c49a4db3ea19048c00f9fff457eac20baa3ff7677" exitCode=0 Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.234357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerDied","Data":"39f3bd4df3b33bcdda9be40c49a4db3ea19048c00f9fff457eac20baa3ff7677"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.240384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" event={"ID":"fe1698aa-b5c6-4480-a886-c89481d08c59","Type":"ContainerDied","Data":"ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.240416 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad4f036bd4b8ac8562f7c115de6a3054624d8518772f724710baf61ef353e602" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.240420 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566742-wqk8p" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.243097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerStarted","Data":"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.246195 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerStarted","Data":"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.247640 4846 generic.go:334] "Generic (PLEG): container finished" podID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" containerID="4bfa365c6112a2b52c8b380c45beec1408fd692036f0d3fd40c624ec3b03310e" exitCode=0 Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.247688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566740-46wjt" event={"ID":"8dd31859-c683-4f1a-b167-3f20a96b6f9c","Type":"ContainerDied","Data":"4bfa365c6112a2b52c8b380c45beec1408fd692036f0d3fd40c624ec3b03310e"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.249297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerStarted","Data":"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1"} Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.279190 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lh8ff" podStartSLOduration=3.305971337 podStartE2EDuration="1m2.279173791s" podCreationTimestamp="2026-03-20 11:01:04 +0000 UTC" firstStartedPulling="2026-03-20 11:01:06.369148237 +0000 UTC m=+237.735556473" lastFinishedPulling="2026-03-20 11:02:05.342350691 +0000 UTC m=+296.708758927" observedRunningTime="2026-03-20 11:02:06.276199161 +0000 UTC m=+297.642607397" watchObservedRunningTime="2026-03-20 11:02:06.279173791 +0000 UTC m=+297.645582037" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.330629 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4z7gv" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="registry-server" probeResult="failure" output=< Mar 20 11:02:06 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:02:06 crc kubenswrapper[4846]: > Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.346866 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6qggw" podStartSLOduration=3.320803363 podStartE2EDuration="1m0.34684802s" podCreationTimestamp="2026-03-20 11:01:06 +0000 UTC" firstStartedPulling="2026-03-20 11:01:08.673503468 +0000 UTC m=+240.039911704" lastFinishedPulling="2026-03-20 11:02:05.699548125 +0000 UTC m=+297.065956361" observedRunningTime="2026-03-20 11:02:06.343477939 +0000 UTC m=+297.709886175" watchObservedRunningTime="2026-03-20 11:02:06.34684802 +0000 UTC m=+297.713256256" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.684626 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:06 crc kubenswrapper[4846]: I0320 11:02:06.685123 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.260021 4846 generic.go:334] "Generic (PLEG): container finished" podID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerID="7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1" exitCode=0 Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.260092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerDied","Data":"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1"} Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.262623 4846 generic.go:334] "Generic (PLEG): container finished" podID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerID="3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210" exitCode=0 Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.262671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerDied","Data":"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210"} Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.308457 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.308808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.603273 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.658413 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9whd\" (UniqueName: \"kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd\") pod \"8dd31859-c683-4f1a-b167-3f20a96b6f9c\" (UID: \"8dd31859-c683-4f1a-b167-3f20a96b6f9c\") " Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.665967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd" (OuterVolumeSpecName: "kube-api-access-s9whd") pod "8dd31859-c683-4f1a-b167-3f20a96b6f9c" (UID: "8dd31859-c683-4f1a-b167-3f20a96b6f9c"). InnerVolumeSpecName "kube-api-access-s9whd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.736014 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6qggw" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="registry-server" probeResult="failure" output=< Mar 20 11:02:07 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:02:07 crc kubenswrapper[4846]: > Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.760478 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9whd\" (UniqueName: \"kubernetes.io/projected/8dd31859-c683-4f1a-b167-3f20a96b6f9c-kube-api-access-s9whd\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.795464 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.795541 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:07 crc kubenswrapper[4846]: I0320 11:02:07.866480 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:08 crc kubenswrapper[4846]: I0320 11:02:08.272201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566740-46wjt" event={"ID":"8dd31859-c683-4f1a-b167-3f20a96b6f9c","Type":"ContainerDied","Data":"79cbc8a2bac7be21f4034f30a0c3e022fef5875c9ff2c518852bca604f266f33"} Mar 20 11:02:08 crc kubenswrapper[4846]: I0320 11:02:08.272736 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79cbc8a2bac7be21f4034f30a0c3e022fef5875c9ff2c518852bca604f266f33" Mar 20 11:02:08 crc kubenswrapper[4846]: I0320 11:02:08.272296 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566740-46wjt" Mar 20 11:02:08 crc kubenswrapper[4846]: I0320 11:02:08.313685 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:08 crc kubenswrapper[4846]: I0320 11:02:08.356401 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kh2d5" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="registry-server" probeResult="failure" output=< Mar 20 11:02:08 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:02:08 crc kubenswrapper[4846]: > Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.285953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerStarted","Data":"654e75bc0015d51a0a4f8cd4fc259f6c082039319f43ee1f2e05d520b54a8d47"} Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.677677 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.678251 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.678314 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.678962 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:02:09 crc kubenswrapper[4846]: I0320 11:02:09.679031 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e" gracePeriod=600 Mar 20 11:02:10 crc kubenswrapper[4846]: I0320 11:02:10.315613 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6vwj9" podStartSLOduration=4.090298214 podStartE2EDuration="1m5.315588846s" podCreationTimestamp="2026-03-20 11:01:05 +0000 UTC" firstStartedPulling="2026-03-20 11:01:07.583049176 +0000 UTC m=+238.949457412" lastFinishedPulling="2026-03-20 11:02:08.808339808 +0000 UTC m=+300.174748044" observedRunningTime="2026-03-20 11:02:10.31352575 +0000 UTC m=+301.679933986" watchObservedRunningTime="2026-03-20 11:02:10.315588846 +0000 UTC m=+301.681997082" Mar 20 11:02:11 crc kubenswrapper[4846]: I0320 11:02:11.308039 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e" exitCode=0 Mar 20 11:02:11 crc kubenswrapper[4846]: I0320 11:02:11.308105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e"} Mar 20 11:02:11 crc kubenswrapper[4846]: I0320 11:02:11.629502 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:02:11 crc kubenswrapper[4846]: I0320 11:02:11.630232 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nhrbq" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="registry-server" containerID="cri-o://95fbc174cd3771e5ed2f271d26205b0ab38e476459b88aed13572bf7a26c2cc3" gracePeriod=2 Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.328790 4846 generic.go:334] "Generic (PLEG): container finished" podID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerID="95fbc174cd3771e5ed2f271d26205b0ab38e476459b88aed13572bf7a26c2cc3" exitCode=0 Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.349924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerDied","Data":"95fbc174cd3771e5ed2f271d26205b0ab38e476459b88aed13572bf7a26c2cc3"} Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.654279 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.857232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities\") pod \"f404899d-1c18-424f-ab1c-2c7c96c2e368\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.857408 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9pfz\" (UniqueName: \"kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz\") pod \"f404899d-1c18-424f-ab1c-2c7c96c2e368\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.857519 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content\") pod \"f404899d-1c18-424f-ab1c-2c7c96c2e368\" (UID: \"f404899d-1c18-424f-ab1c-2c7c96c2e368\") " Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.859865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities" (OuterVolumeSpecName: "utilities") pod "f404899d-1c18-424f-ab1c-2c7c96c2e368" (UID: "f404899d-1c18-424f-ab1c-2c7c96c2e368"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.867067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz" (OuterVolumeSpecName: "kube-api-access-r9pfz") pod "f404899d-1c18-424f-ab1c-2c7c96c2e368" (UID: "f404899d-1c18-424f-ab1c-2c7c96c2e368"). InnerVolumeSpecName "kube-api-access-r9pfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.959867 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:13 crc kubenswrapper[4846]: I0320 11:02:13.959926 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9pfz\" (UniqueName: \"kubernetes.io/projected/f404899d-1c18-424f-ab1c-2c7c96c2e368-kube-api-access-r9pfz\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.008070 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f404899d-1c18-424f-ab1c-2c7c96c2e368" (UID: "f404899d-1c18-424f-ab1c-2c7c96c2e368"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.061350 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f404899d-1c18-424f-ab1c-2c7c96c2e368-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.341273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nhrbq" event={"ID":"f404899d-1c18-424f-ab1c-2c7c96c2e368","Type":"ContainerDied","Data":"120c84aaf198f9cc55ea03048a64e1603bae415f39dc2e7bfe920ce69e5e0ed9"} Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.342002 4846 scope.go:117] "RemoveContainer" containerID="95fbc174cd3771e5ed2f271d26205b0ab38e476459b88aed13572bf7a26c2cc3" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.341348 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nhrbq" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.368182 4846 scope.go:117] "RemoveContainer" containerID="11c3f05d48489567e31bbe3dd4ab70142418e8d41c983b4986c99498f7a8fb4e" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.392489 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.397378 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nhrbq"] Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.404467 4846 scope.go:117] "RemoveContainer" containerID="78310e72e3d3c96b337ca94613a7e79f387a9afa753caafcbf85328ea54a708d" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.556288 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.556361 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.610756 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.788561 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:14 crc kubenswrapper[4846]: I0320 11:02:14.834867 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.331246 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" path="/var/lib/kubelet/pods/f404899d-1c18-424f-ab1c-2c7c96c2e368/volumes" Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.350401 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerStarted","Data":"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198"} Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.354404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerStarted","Data":"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c"} Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.357667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c"} Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.386691 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt9fr" podStartSLOduration=4.9113787 podStartE2EDuration="1m12.386665216s" podCreationTimestamp="2026-03-20 11:01:03 +0000 UTC" firstStartedPulling="2026-03-20 11:01:06.351977072 +0000 UTC m=+237.718385308" lastFinishedPulling="2026-03-20 11:02:13.827263588 +0000 UTC m=+305.193671824" observedRunningTime="2026-03-20 11:02:15.386014768 +0000 UTC m=+306.752423004" watchObservedRunningTime="2026-03-20 11:02:15.386665216 +0000 UTC m=+306.753073782" Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.421652 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:15 crc kubenswrapper[4846]: I0320 11:02:15.431521 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-56jhc" podStartSLOduration=6.18044181 podStartE2EDuration="1m12.431490997s" podCreationTimestamp="2026-03-20 11:01:03 +0000 UTC" firstStartedPulling="2026-03-20 11:01:05.370826945 +0000 UTC m=+236.737235181" lastFinishedPulling="2026-03-20 11:02:11.621876082 +0000 UTC m=+302.988284368" observedRunningTime="2026-03-20 11:02:15.429434181 +0000 UTC m=+306.795842417" watchObservedRunningTime="2026-03-20 11:02:15.431490997 +0000 UTC m=+306.797899233" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.272336 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.272438 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.347038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.417015 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.744439 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:16 crc kubenswrapper[4846]: I0320 11:02:16.814256 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.027769 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.371388 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lh8ff" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="registry-server" containerID="cri-o://daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839" gracePeriod=2 Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.384103 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.437990 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.909288 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.948062 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgwxs\" (UniqueName: \"kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs\") pod \"408ee9c3-1b20-4002-850f-328b27b972d0\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.948150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities\") pod \"408ee9c3-1b20-4002-850f-328b27b972d0\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.948248 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content\") pod \"408ee9c3-1b20-4002-850f-328b27b972d0\" (UID: \"408ee9c3-1b20-4002-850f-328b27b972d0\") " Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.951696 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities" (OuterVolumeSpecName: "utilities") pod "408ee9c3-1b20-4002-850f-328b27b972d0" (UID: "408ee9c3-1b20-4002-850f-328b27b972d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:17 crc kubenswrapper[4846]: I0320 11:02:17.958671 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs" (OuterVolumeSpecName: "kube-api-access-qgwxs") pod "408ee9c3-1b20-4002-850f-328b27b972d0" (UID: "408ee9c3-1b20-4002-850f-328b27b972d0"). InnerVolumeSpecName "kube-api-access-qgwxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.023581 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.024683 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4z7gv" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="registry-server" containerID="cri-o://61386f7d4a6aef54983984e41c8e91c231429f0f56c6fee7bedfd72caa6f0395" gracePeriod=2 Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.042354 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "408ee9c3-1b20-4002-850f-328b27b972d0" (UID: "408ee9c3-1b20-4002-850f-328b27b972d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.049496 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.049542 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgwxs\" (UniqueName: \"kubernetes.io/projected/408ee9c3-1b20-4002-850f-328b27b972d0-kube-api-access-qgwxs\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.049559 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408ee9c3-1b20-4002-850f-328b27b972d0-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.386300 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ba6d606-56e4-4285-887a-4b86f8198046" containerID="61386f7d4a6aef54983984e41c8e91c231429f0f56c6fee7bedfd72caa6f0395" exitCode=0 Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.386392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerDied","Data":"61386f7d4a6aef54983984e41c8e91c231429f0f56c6fee7bedfd72caa6f0395"} Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.389697 4846 generic.go:334] "Generic (PLEG): container finished" podID="408ee9c3-1b20-4002-850f-328b27b972d0" containerID="daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839" exitCode=0 Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.390791 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lh8ff" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.392053 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerDied","Data":"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839"} Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.392108 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lh8ff" event={"ID":"408ee9c3-1b20-4002-850f-328b27b972d0","Type":"ContainerDied","Data":"8105e2e50c571fe2eb64207670da80cd57289abd8a3a2f5cf97fa65fa323adb7"} Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.392133 4846 scope.go:117] "RemoveContainer" containerID="daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.425167 4846 scope.go:117] "RemoveContainer" containerID="971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.434399 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.438284 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lh8ff"] Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.446862 4846 scope.go:117] "RemoveContainer" containerID="8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.470311 4846 scope.go:117] "RemoveContainer" containerID="daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839" Mar 20 11:02:18 crc kubenswrapper[4846]: E0320 11:02:18.471150 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839\": container with ID starting with daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839 not found: ID does not exist" containerID="daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.471230 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839"} err="failed to get container status \"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839\": rpc error: code = NotFound desc = could not find container \"daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839\": container with ID starting with daa47cf6cc88ce9d80ee4f8ed79f6521d41585d141f26951351475afe035d839 not found: ID does not exist" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.471277 4846 scope.go:117] "RemoveContainer" containerID="971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe" Mar 20 11:02:18 crc kubenswrapper[4846]: E0320 11:02:18.471833 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe\": container with ID starting with 971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe not found: ID does not exist" containerID="971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.471891 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe"} err="failed to get container status \"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe\": rpc error: code = NotFound desc = could not find container \"971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe\": container with ID starting with 971734b1a7a9f90c6f77bc9095072127011d1240ce3cea303ffebb2a12d65bbe not found: ID does not exist" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.471957 4846 scope.go:117] "RemoveContainer" containerID="8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b" Mar 20 11:02:18 crc kubenswrapper[4846]: E0320 11:02:18.472315 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b\": container with ID starting with 8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b not found: ID does not exist" containerID="8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.472343 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b"} err="failed to get container status \"8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b\": rpc error: code = NotFound desc = could not find container \"8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b\": container with ID starting with 8eeb83b6cb50c5142d4762c31617e701f6578af750ee1c5219dc7d286fee595b not found: ID does not exist" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.559694 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.658717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content\") pod \"6ba6d606-56e4-4285-887a-4b86f8198046\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.658960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities\") pod \"6ba6d606-56e4-4285-887a-4b86f8198046\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.659013 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7dfm\" (UniqueName: \"kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm\") pod \"6ba6d606-56e4-4285-887a-4b86f8198046\" (UID: \"6ba6d606-56e4-4285-887a-4b86f8198046\") " Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.660007 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities" (OuterVolumeSpecName: "utilities") pod "6ba6d606-56e4-4285-887a-4b86f8198046" (UID: "6ba6d606-56e4-4285-887a-4b86f8198046"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.666325 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm" (OuterVolumeSpecName: "kube-api-access-q7dfm") pod "6ba6d606-56e4-4285-887a-4b86f8198046" (UID: "6ba6d606-56e4-4285-887a-4b86f8198046"). InnerVolumeSpecName "kube-api-access-q7dfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.708025 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ba6d606-56e4-4285-887a-4b86f8198046" (UID: "6ba6d606-56e4-4285-887a-4b86f8198046"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.760161 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.760208 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7dfm\" (UniqueName: \"kubernetes.io/projected/6ba6d606-56e4-4285-887a-4b86f8198046-kube-api-access-q7dfm\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:18 crc kubenswrapper[4846]: I0320 11:02:18.760222 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba6d606-56e4-4285-887a-4b86f8198046-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.337661 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" path="/var/lib/kubelet/pods/408ee9c3-1b20-4002-850f-328b27b972d0/volumes" Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.403600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4z7gv" event={"ID":"6ba6d606-56e4-4285-887a-4b86f8198046","Type":"ContainerDied","Data":"75018c7f04688ad7561a3f230f6bbef978b7678fb85c43ca20f8bfc3e8dd6ae0"} Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.403700 4846 scope.go:117] "RemoveContainer" containerID="61386f7d4a6aef54983984e41c8e91c231429f0f56c6fee7bedfd72caa6f0395" Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.404232 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4z7gv" Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.439105 4846 scope.go:117] "RemoveContainer" containerID="95dad7bc2edb7648d231390bb4815733ea47c81d9e138016e60003b804692b61" Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.441558 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.448100 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4z7gv"] Mar 20 11:02:19 crc kubenswrapper[4846]: I0320 11:02:19.467694 4846 scope.go:117] "RemoveContainer" containerID="2b6a518874b790be561c435c3114c0df984e1e56213da3d0baffc76761dc1335" Mar 20 11:02:20 crc kubenswrapper[4846]: I0320 11:02:20.421366 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:02:20 crc kubenswrapper[4846]: I0320 11:02:20.421728 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6qggw" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="registry-server" containerID="cri-o://8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57" gracePeriod=2 Mar 20 11:02:20 crc kubenswrapper[4846]: I0320 11:02:20.937529 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.006986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content\") pod \"131db4bd-a795-401c-a9e7-b66f4512dc32\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.007420 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mq5x\" (UniqueName: \"kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x\") pod \"131db4bd-a795-401c-a9e7-b66f4512dc32\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.007552 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities\") pod \"131db4bd-a795-401c-a9e7-b66f4512dc32\" (UID: \"131db4bd-a795-401c-a9e7-b66f4512dc32\") " Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.008662 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities" (OuterVolumeSpecName: "utilities") pod "131db4bd-a795-401c-a9e7-b66f4512dc32" (UID: "131db4bd-a795-401c-a9e7-b66f4512dc32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.013095 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x" (OuterVolumeSpecName: "kube-api-access-5mq5x") pod "131db4bd-a795-401c-a9e7-b66f4512dc32" (UID: "131db4bd-a795-401c-a9e7-b66f4512dc32"). InnerVolumeSpecName "kube-api-access-5mq5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.069136 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "131db4bd-a795-401c-a9e7-b66f4512dc32" (UID: "131db4bd-a795-401c-a9e7-b66f4512dc32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.109679 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.109729 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mq5x\" (UniqueName: \"kubernetes.io/projected/131db4bd-a795-401c-a9e7-b66f4512dc32-kube-api-access-5mq5x\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.109747 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131db4bd-a795-401c-a9e7-b66f4512dc32-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.337653 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" path="/var/lib/kubelet/pods/6ba6d606-56e4-4285-887a-4b86f8198046/volumes" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.430622 4846 generic.go:334] "Generic (PLEG): container finished" podID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerID="8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57" exitCode=0 Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.430725 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerDied","Data":"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57"} Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.431268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6qggw" event={"ID":"131db4bd-a795-401c-a9e7-b66f4512dc32","Type":"ContainerDied","Data":"7416ad4f63b2f88eecdd33df72960cd715dd51b0f92b9bf9f51d4f8be9ed7f52"} Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.431333 4846 scope.go:117] "RemoveContainer" containerID="8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.430828 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6qggw" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.462559 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.466390 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6qggw"] Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.467187 4846 scope.go:117] "RemoveContainer" containerID="810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.486112 4846 scope.go:117] "RemoveContainer" containerID="6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.509610 4846 scope.go:117] "RemoveContainer" containerID="8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57" Mar 20 11:02:21 crc kubenswrapper[4846]: E0320 11:02:21.510365 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57\": container with ID starting with 8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57 not found: ID does not exist" containerID="8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.510422 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57"} err="failed to get container status \"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57\": rpc error: code = NotFound desc = could not find container \"8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57\": container with ID starting with 8e05e462cdc5f010b84caae3dc330706370124664ec11567053cdc20c3006b57 not found: ID does not exist" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.510459 4846 scope.go:117] "RemoveContainer" containerID="810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8" Mar 20 11:02:21 crc kubenswrapper[4846]: E0320 11:02:21.511029 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8\": container with ID starting with 810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8 not found: ID does not exist" containerID="810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.511069 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8"} err="failed to get container status \"810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8\": rpc error: code = NotFound desc = could not find container \"810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8\": container with ID starting with 810aa571151ce60401f93587b71add578d1ab6a35d1f0259160a1515c2afddb8 not found: ID does not exist" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.511094 4846 scope.go:117] "RemoveContainer" containerID="6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834" Mar 20 11:02:21 crc kubenswrapper[4846]: E0320 11:02:21.511548 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834\": container with ID starting with 6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834 not found: ID does not exist" containerID="6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834" Mar 20 11:02:21 crc kubenswrapper[4846]: I0320 11:02:21.511576 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834"} err="failed to get container status \"6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834\": rpc error: code = NotFound desc = could not find container \"6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834\": container with ID starting with 6e116f098f62e501256c331747533827812c00778e8a9bfed0809f283c24a834 not found: ID does not exist" Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.318457 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.319304 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" podUID="875040a8-922a-4658-b3ec-205a1341d541" containerName="controller-manager" containerID="cri-o://2022d6a1102d6c110d784b753fe69b40d8116d69e5a4e6e52e5206797c3251fd" gracePeriod=30 Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.418458 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.418694 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" podUID="bc42705d-a520-4db2-994b-c626d1ed2e34" containerName="route-controller-manager" containerID="cri-o://f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c" gracePeriod=30 Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.455573 4846 generic.go:334] "Generic (PLEG): container finished" podID="875040a8-922a-4658-b3ec-205a1341d541" containerID="2022d6a1102d6c110d784b753fe69b40d8116d69e5a4e6e52e5206797c3251fd" exitCode=0 Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.455670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" event={"ID":"875040a8-922a-4658-b3ec-205a1341d541","Type":"ContainerDied","Data":"2022d6a1102d6c110d784b753fe69b40d8116d69e5a4e6e52e5206797c3251fd"} Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.964928 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:22 crc kubenswrapper[4846]: I0320 11:02:22.972355 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.038846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles\") pod \"875040a8-922a-4658-b3ec-205a1341d541\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.038938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert\") pod \"875040a8-922a-4658-b3ec-205a1341d541\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.038959 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert\") pod \"bc42705d-a520-4db2-994b-c626d1ed2e34\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.038983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62754\" (UniqueName: \"kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754\") pod \"875040a8-922a-4658-b3ec-205a1341d541\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.039006 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config\") pod \"875040a8-922a-4658-b3ec-205a1341d541\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.039077 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwg4f\" (UniqueName: \"kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f\") pod \"bc42705d-a520-4db2-994b-c626d1ed2e34\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.039095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca\") pod \"bc42705d-a520-4db2-994b-c626d1ed2e34\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.039113 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config\") pod \"bc42705d-a520-4db2-994b-c626d1ed2e34\" (UID: \"bc42705d-a520-4db2-994b-c626d1ed2e34\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.039143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca\") pod \"875040a8-922a-4658-b3ec-205a1341d541\" (UID: \"875040a8-922a-4658-b3ec-205a1341d541\") " Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.040203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config" (OuterVolumeSpecName: "config") pod "875040a8-922a-4658-b3ec-205a1341d541" (UID: "875040a8-922a-4658-b3ec-205a1341d541"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.040238 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca" (OuterVolumeSpecName: "client-ca") pod "875040a8-922a-4658-b3ec-205a1341d541" (UID: "875040a8-922a-4658-b3ec-205a1341d541"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.040419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "875040a8-922a-4658-b3ec-205a1341d541" (UID: "875040a8-922a-4658-b3ec-205a1341d541"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.041477 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca" (OuterVolumeSpecName: "client-ca") pod "bc42705d-a520-4db2-994b-c626d1ed2e34" (UID: "bc42705d-a520-4db2-994b-c626d1ed2e34"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.042407 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config" (OuterVolumeSpecName: "config") pod "bc42705d-a520-4db2-994b-c626d1ed2e34" (UID: "bc42705d-a520-4db2-994b-c626d1ed2e34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.046563 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f" (OuterVolumeSpecName: "kube-api-access-cwg4f") pod "bc42705d-a520-4db2-994b-c626d1ed2e34" (UID: "bc42705d-a520-4db2-994b-c626d1ed2e34"). InnerVolumeSpecName "kube-api-access-cwg4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.047096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "875040a8-922a-4658-b3ec-205a1341d541" (UID: "875040a8-922a-4658-b3ec-205a1341d541"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.047826 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc42705d-a520-4db2-994b-c626d1ed2e34" (UID: "bc42705d-a520-4db2-994b-c626d1ed2e34"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.051028 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754" (OuterVolumeSpecName: "kube-api-access-62754") pod "875040a8-922a-4658-b3ec-205a1341d541" (UID: "875040a8-922a-4658-b3ec-205a1341d541"). InnerVolumeSpecName "kube-api-access-62754". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141446 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141492 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141506 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875040a8-922a-4658-b3ec-205a1341d541-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141519 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc42705d-a520-4db2-994b-c626d1ed2e34-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141530 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875040a8-922a-4658-b3ec-205a1341d541-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141541 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62754\" (UniqueName: \"kubernetes.io/projected/875040a8-922a-4658-b3ec-205a1341d541-kube-api-access-62754\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141551 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwg4f\" (UniqueName: \"kubernetes.io/projected/bc42705d-a520-4db2-994b-c626d1ed2e34-kube-api-access-cwg4f\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141561 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.141571 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc42705d-a520-4db2-994b-c626d1ed2e34-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.329227 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" path="/var/lib/kubelet/pods/131db4bd-a795-401c-a9e7-b66f4512dc32/volumes" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.466940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" event={"ID":"875040a8-922a-4658-b3ec-205a1341d541","Type":"ContainerDied","Data":"aa4aeb22c6c968d4a51d222905a69295c07e39a189f1b0b4e07aa2b322f2d8d4"} Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.467341 4846 scope.go:117] "RemoveContainer" containerID="2022d6a1102d6c110d784b753fe69b40d8116d69e5a4e6e52e5206797c3251fd" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.466964 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-679bd75db9-w7fqd" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.469069 4846 generic.go:334] "Generic (PLEG): container finished" podID="bc42705d-a520-4db2-994b-c626d1ed2e34" containerID="f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c" exitCode=0 Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.469121 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.469145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" event={"ID":"bc42705d-a520-4db2-994b-c626d1ed2e34","Type":"ContainerDied","Data":"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c"} Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.469460 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z" event={"ID":"bc42705d-a520-4db2-994b-c626d1ed2e34","Type":"ContainerDied","Data":"87a1910f5cb4e0866fa602a00d6175d5460033f002f9c868284d2454498efc5c"} Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.490561 4846 scope.go:117] "RemoveContainer" containerID="f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.497013 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.500345 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bcf5d65c9-sxx5z"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.511940 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.517191 4846 scope.go:117] "RemoveContainer" containerID="f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.518501 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c\": container with ID starting with f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c not found: ID does not exist" containerID="f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.518580 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c"} err="failed to get container status \"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c\": rpc error: code = NotFound desc = could not find container \"f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c\": container with ID starting with f4aad90ba1e4f5f807c925f5b9b2e72bb343eff88a6c46037bba1b89a9cbec2c not found: ID does not exist" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.519655 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs"] Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520055 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875040a8-922a-4658-b3ec-205a1341d541" containerName="controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520082 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="875040a8-922a-4658-b3ec-205a1341d541" containerName="controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520110 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520122 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520143 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc42705d-a520-4db2-994b-c626d1ed2e34" containerName="route-controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520152 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc42705d-a520-4db2-994b-c626d1ed2e34" containerName="route-controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520166 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520175 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520188 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520201 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520212 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520224 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520239 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520249 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520264 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520275 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520288 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1698aa-b5c6-4480-a886-c89481d08c59" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520296 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1698aa-b5c6-4480-a886-c89481d08c59" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520309 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520319 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520337 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520347 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="extract-content" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520357 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520366 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520382 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520390 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520400 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520410 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="extract-utilities" Mar 20 11:02:23 crc kubenswrapper[4846]: E0320 11:02:23.520421 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520430 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520568 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="131db4bd-a795-401c-a9e7-b66f4512dc32" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520583 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1698aa-b5c6-4480-a886-c89481d08c59" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520595 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" containerName="oc" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520611 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f404899d-1c18-424f-ab1c-2c7c96c2e368" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520628 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="875040a8-922a-4658-b3ec-205a1341d541" containerName="controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520670 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="408ee9c3-1b20-4002-850f-328b27b972d0" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520682 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba6d606-56e4-4285-887a-4b86f8198046" containerName="registry-server" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.520691 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc42705d-a520-4db2-994b-c626d1ed2e34" containerName="route-controller-manager" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.521283 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524224 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524324 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524451 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524608 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524736 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.524799 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.526812 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6864c76f77-c7h7l"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.527939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.530140 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.530596 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.530730 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.531011 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.531294 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.532285 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.535163 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-679bd75db9-w7fqd"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.538846 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.542270 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6864c76f77-c7h7l"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.545574 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs"] Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648529 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-config\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648598 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5w7q\" (UniqueName: \"kubernetes.io/projected/c3f46ad3-2d2d-4148-adb8-4e4798420da7-kube-api-access-g5w7q\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3f46ad3-2d2d-4148-adb8-4e4798420da7-serving-cert\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-proxy-ca-bundles\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-client-ca\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.648956 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-serving-cert\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.649089 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpmg8\" (UniqueName: \"kubernetes.io/projected/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-kube-api-access-vpmg8\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.649149 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-client-ca\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.649192 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-config\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-serving-cert\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpmg8\" (UniqueName: \"kubernetes.io/projected/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-kube-api-access-vpmg8\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-client-ca\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750141 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-config\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750188 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-config\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750216 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5w7q\" (UniqueName: \"kubernetes.io/projected/c3f46ad3-2d2d-4148-adb8-4e4798420da7-kube-api-access-g5w7q\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3f46ad3-2d2d-4148-adb8-4e4798420da7-serving-cert\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-proxy-ca-bundles\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.750280 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-client-ca\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.751491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-client-ca\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.751942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-config\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.752128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-client-ca\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.754509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-proxy-ca-bundles\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.754780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3f46ad3-2d2d-4148-adb8-4e4798420da7-config\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.757726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3f46ad3-2d2d-4148-adb8-4e4798420da7-serving-cert\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.763541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-serving-cert\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.772355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpmg8\" (UniqueName: \"kubernetes.io/projected/a662c2d2-c060-4ff4-a529-9e2a3aaff0e2-kube-api-access-vpmg8\") pod \"route-controller-manager-59dcdc8755-thtvs\" (UID: \"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2\") " pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.791021 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5w7q\" (UniqueName: \"kubernetes.io/projected/c3f46ad3-2d2d-4148-adb8-4e4798420da7-kube-api-access-g5w7q\") pod \"controller-manager-6864c76f77-c7h7l\" (UID: \"c3f46ad3-2d2d-4148-adb8-4e4798420da7\") " pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.845368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:23 crc kubenswrapper[4846]: I0320 11:02:23.858477 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.063369 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.063444 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.121356 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.191169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6864c76f77-c7h7l"] Mar 20 11:02:24 crc kubenswrapper[4846]: W0320 11:02:24.199351 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3f46ad3_2d2d_4148_adb8_4e4798420da7.slice/crio-27b0ae784033f51c575d87681a0cda1214d0b6d997f798b27fd639d13167d4a4 WatchSource:0}: Error finding container 27b0ae784033f51c575d87681a0cda1214d0b6d997f798b27fd639d13167d4a4: Status 404 returned error can't find the container with id 27b0ae784033f51c575d87681a0cda1214d0b6d997f798b27fd639d13167d4a4 Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.334083 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs"] Mar 20 11:02:24 crc kubenswrapper[4846]: W0320 11:02:24.340344 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda662c2d2_c060_4ff4_a529_9e2a3aaff0e2.slice/crio-1f3d4ed1a38126bfd9d43d3fef9f4e503c298df2eb5725d96e5d957227e7e7a8 WatchSource:0}: Error finding container 1f3d4ed1a38126bfd9d43d3fef9f4e503c298df2eb5725d96e5d957227e7e7a8: Status 404 returned error can't find the container with id 1f3d4ed1a38126bfd9d43d3fef9f4e503c298df2eb5725d96e5d957227e7e7a8 Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.480759 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" event={"ID":"c3f46ad3-2d2d-4148-adb8-4e4798420da7","Type":"ContainerStarted","Data":"963505e36eaa9344a4c972ff81deceb0c15f6d9d9e3611e7cf8b39347d527679"} Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.480838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" event={"ID":"c3f46ad3-2d2d-4148-adb8-4e4798420da7","Type":"ContainerStarted","Data":"27b0ae784033f51c575d87681a0cda1214d0b6d997f798b27fd639d13167d4a4"} Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.481053 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.483044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" event={"ID":"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2","Type":"ContainerStarted","Data":"238e9d53dca1c5ab1f69142a065ec9d365eb9930f8cd2a0c94bf1b995e6db845"} Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.483112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" event={"ID":"a662c2d2-c060-4ff4-a529-9e2a3aaff0e2","Type":"ContainerStarted","Data":"1f3d4ed1a38126bfd9d43d3fef9f4e503c298df2eb5725d96e5d957227e7e7a8"} Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.483134 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.484714 4846 patch_prober.go:28] interesting pod/route-controller-manager-59dcdc8755-thtvs container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.484778 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" podUID="a662c2d2-c060-4ff4-a529-9e2a3aaff0e2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.488059 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.509690 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6864c76f77-c7h7l" podStartSLOduration=2.509660819 podStartE2EDuration="2.509660819s" podCreationTimestamp="2026-03-20 11:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:02:24.505514217 +0000 UTC m=+315.871922483" watchObservedRunningTime="2026-03-20 11:02:24.509660819 +0000 UTC m=+315.876069075" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.543178 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.552760 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" podStartSLOduration=2.5527361219999998 podStartE2EDuration="2.552736122s" podCreationTimestamp="2026-03-20 11:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:02:24.549761532 +0000 UTC m=+315.916169768" watchObservedRunningTime="2026-03-20 11:02:24.552736122 +0000 UTC m=+315.919144358" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.593391 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.593481 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:02:24 crc kubenswrapper[4846]: I0320 11:02:24.650618 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:02:25 crc kubenswrapper[4846]: I0320 11:02:25.330617 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875040a8-922a-4658-b3ec-205a1341d541" path="/var/lib/kubelet/pods/875040a8-922a-4658-b3ec-205a1341d541/volumes" Mar 20 11:02:25 crc kubenswrapper[4846]: I0320 11:02:25.331205 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc42705d-a520-4db2-994b-c626d1ed2e34" path="/var/lib/kubelet/pods/bc42705d-a520-4db2-994b-c626d1ed2e34/volumes" Mar 20 11:02:25 crc kubenswrapper[4846]: I0320 11:02:25.498341 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59dcdc8755-thtvs" Mar 20 11:02:25 crc kubenswrapper[4846]: I0320 11:02:25.555105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:02:25 crc kubenswrapper[4846]: I0320 11:02:25.756807 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kgqq8"] Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.088771 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.090432 4846 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.090652 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.090935 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71" gracePeriod=15 Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.090988 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.090995 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d" gracePeriod=15 Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.091121 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4" gracePeriod=15 Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.091315 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846" gracePeriod=15 Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.091821 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.091879 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092091 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092100 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092114 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092121 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092134 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092142 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092205 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092212 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092222 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092229 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092239 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092245 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092260 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092268 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092382 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092394 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092402 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092411 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092421 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092431 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092441 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092546 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092554 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092646 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: E0320 11:02:29.092740 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092748 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.092863 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.091818 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e" gracePeriod=15 Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.106805 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.179754 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.281440 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.281980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282022 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.281638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282169 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282594 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.282737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.525587 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.527352 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 11:02:29 crc kubenswrapper[4846]: I0320 11:02:29.528576 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e" exitCode=2 Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.537611 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.539436 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.540190 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4" exitCode=0 Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.540230 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71" exitCode=0 Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.540239 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846" exitCode=0 Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.540333 4846 scope.go:117] "RemoveContainer" containerID="923d37fc6a59d8514b28e0905d375f1dfa7c7a59bed856fa53183f13dbbdc37e" Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.542658 4846 generic.go:334] "Generic (PLEG): container finished" podID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" containerID="b1eccb669b422d66113b2a4ed1138d23c4122eabf9098002c4ae503e15b25e0c" exitCode=0 Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.542714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3aa37f4d-5b65-45f4-ae84-a1f563508ff6","Type":"ContainerDied","Data":"b1eccb669b422d66113b2a4ed1138d23c4122eabf9098002c4ae503e15b25e0c"} Mar 20 11:02:30 crc kubenswrapper[4846]: I0320 11:02:30.543789 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.552680 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.554367 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d" exitCode=0 Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.554428 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be28cdd04ba0c0c312fc8bd425a2c37f263e1b5f584107133af0a929c5ee79c8" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.589152 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.590231 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.590761 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.590961 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.723550 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.723754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.723969 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.724085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.724086 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.725436 4846 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.725484 4846 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.725565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.827083 4846 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.934211 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.935639 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:31 crc kubenswrapper[4846]: I0320 11:02:31.936406 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.032535 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock\") pod \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.032615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access\") pod \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.032670 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir\") pod \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\" (UID: \"3aa37f4d-5b65-45f4-ae84-a1f563508ff6\") " Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.032992 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3aa37f4d-5b65-45f4-ae84-a1f563508ff6" (UID: "3aa37f4d-5b65-45f4-ae84-a1f563508ff6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.033039 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock" (OuterVolumeSpecName: "var-lock") pod "3aa37f4d-5b65-45f4-ae84-a1f563508ff6" (UID: "3aa37f4d-5b65-45f4-ae84-a1f563508ff6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.040250 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3aa37f4d-5b65-45f4-ae84-a1f563508ff6" (UID: "3aa37f4d-5b65-45f4-ae84-a1f563508ff6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.134049 4846 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.134522 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.134694 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3aa37f4d-5b65-45f4-ae84-a1f563508ff6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.567222 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3aa37f4d-5b65-45f4-ae84-a1f563508ff6","Type":"ContainerDied","Data":"88d2027a148d51d187584d9034c1ddb74e175f33e84b6c5e22802a0f7833b910"} Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.567345 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88d2027a148d51d187584d9034c1ddb74e175f33e84b6c5e22802a0f7833b910" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.567281 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.567273 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.596780 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.597471 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.605695 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:32 crc kubenswrapper[4846]: I0320 11:02:32.606164 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:33 crc kubenswrapper[4846]: I0320 11:02:33.331270 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 20 11:02:34 crc kubenswrapper[4846]: E0320 11:02:34.131946 4846 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.212:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:34 crc kubenswrapper[4846]: I0320 11:02:34.133038 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:34 crc kubenswrapper[4846]: E0320 11:02:34.165501 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.212:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e87be30eadb9b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 11:02:34.164788123 +0000 UTC m=+325.531196359,LastTimestamp:2026-03-20 11:02:34.164788123 +0000 UTC m=+325.531196359,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 11:02:34 crc kubenswrapper[4846]: I0320 11:02:34.584810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c"} Mar 20 11:02:34 crc kubenswrapper[4846]: I0320 11:02:34.585389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"86f78a5f385b26b2bf58ba9d8c5783be60b205ddee71e40fa92ce4d71e6cea4f"} Mar 20 11:02:34 crc kubenswrapper[4846]: I0320 11:02:34.586150 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:34 crc kubenswrapper[4846]: E0320 11:02:34.586310 4846 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.212:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.132481 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.132712 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.132881 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.133060 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.133227 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:35 crc kubenswrapper[4846]: I0320 11:02:35.133252 4846 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.133413 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="200ms" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.333938 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="400ms" Mar 20 11:02:35 crc kubenswrapper[4846]: E0320 11:02:35.736681 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="800ms" Mar 20 11:02:36 crc kubenswrapper[4846]: E0320 11:02:36.538120 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="1.6s" Mar 20 11:02:36 crc kubenswrapper[4846]: E0320 11:02:36.685321 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.212:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189e87be30eadb9b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 11:02:34.164788123 +0000 UTC m=+325.531196359,LastTimestamp:2026-03-20 11:02:34.164788123 +0000 UTC m=+325.531196359,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 11:02:38 crc kubenswrapper[4846]: E0320 11:02:38.139779 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="3.2s" Mar 20 11:02:39 crc kubenswrapper[4846]: I0320 11:02:39.327882 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:39 crc kubenswrapper[4846]: E0320 11:02:39.363288 4846 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.212:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" volumeName="registry-storage" Mar 20 11:02:41 crc kubenswrapper[4846]: E0320 11:02:41.340536 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.212:6443: connect: connection refused" interval="6.4s" Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.648599 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.650647 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.650959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345"} Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.650797 4846 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345" exitCode=1 Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.652378 4846 scope.go:117] "RemoveContainer" containerID="138aeb7e1804f0eac9d1c3476a7616f7ebb97b71d7905af660ddbd0ef0530345" Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.652699 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:42 crc kubenswrapper[4846]: I0320 11:02:42.653491 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.012233 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.322366 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.324711 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.325172 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.346589 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.346639 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:43 crc kubenswrapper[4846]: E0320 11:02:43.347349 4846 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.348196 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:43 crc kubenswrapper[4846]: W0320 11:02:43.380637 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-55421ee1fd3ec62fde27c2234206cf1bd2a483dbdf11b3116a5c3d44e086b484 WatchSource:0}: Error finding container 55421ee1fd3ec62fde27c2234206cf1bd2a483dbdf11b3116a5c3d44e086b484: Status 404 returned error can't find the container with id 55421ee1fd3ec62fde27c2234206cf1bd2a483dbdf11b3116a5c3d44e086b484 Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.665692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"55421ee1fd3ec62fde27c2234206cf1bd2a483dbdf11b3116a5c3d44e086b484"} Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.671755 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.673356 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.673451 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e334c23b31525f788b24655117b8fec8c4c46de4ae0e9da695f66c55f0aa7672"} Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.674756 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:43 crc kubenswrapper[4846]: I0320 11:02:43.675390 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.683788 4846 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="85b80c53943de908c5516049f451cd43c72fcff9104d8dcbfb6871e03a2db9ef" exitCode=0 Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.683868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"85b80c53943de908c5516049f451cd43c72fcff9104d8dcbfb6871e03a2db9ef"} Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.684248 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.684283 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.685056 4846 status_manager.go:851] "Failed to get status for pod" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:44 crc kubenswrapper[4846]: E0320 11:02:44.685134 4846 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:44 crc kubenswrapper[4846]: I0320 11:02:44.685592 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.212:6443: connect: connection refused" Mar 20 11:02:45 crc kubenswrapper[4846]: I0320 11:02:45.696139 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c5929de83297c238f379375a7daec527060552d42c688197f13e02bffe2908b8"} Mar 20 11:02:45 crc kubenswrapper[4846]: I0320 11:02:45.698520 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fefabfb10fc21bef40e0fa552f3c3008cae0298e4743a7cb733c4e247bc6e94b"} Mar 20 11:02:45 crc kubenswrapper[4846]: I0320 11:02:45.698589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d1fd3c1b5b2302f4c7fb83f9df00530bc8935632b8699e0841b0119f86d64cba"} Mar 20 11:02:46 crc kubenswrapper[4846]: I0320 11:02:46.706797 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3546dade97e402a0f9393ac1c78b92cff8944a494a7fa79c5d88b4ebae1cda62"} Mar 20 11:02:46 crc kubenswrapper[4846]: I0320 11:02:46.707284 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:46 crc kubenswrapper[4846]: I0320 11:02:46.707313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c46d8fd3ffd2a8155ad8350d7dfa357f33a33acc95eeaf470ebf8f504f1ebb07"} Mar 20 11:02:46 crc kubenswrapper[4846]: I0320 11:02:46.707334 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:46 crc kubenswrapper[4846]: I0320 11:02:46.707315 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:47 crc kubenswrapper[4846]: I0320 11:02:47.415542 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 11:02:47 crc kubenswrapper[4846]: I0320 11:02:47.426967 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 11:02:47 crc kubenswrapper[4846]: I0320 11:02:47.713130 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 11:02:48 crc kubenswrapper[4846]: I0320 11:02:48.349227 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:48 crc kubenswrapper[4846]: I0320 11:02:48.349296 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:48 crc kubenswrapper[4846]: I0320 11:02:48.354283 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:50 crc kubenswrapper[4846]: I0320 11:02:50.793985 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" containerName="oauth-openshift" containerID="cri-o://352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a" gracePeriod=15 Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.276602 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452204 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452364 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx5ts\" (UniqueName: \"kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452574 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452658 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452706 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452753 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452871 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.452960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.453011 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.453061 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.453114 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.453188 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error\") pod \"4acf8a6b-4d95-435f-a697-49a803be1454\" (UID: \"4acf8a6b-4d95-435f-a697-49a803be1454\") " Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.456773 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.458678 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.468485 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.469381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.468975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts" (OuterVolumeSpecName: "kube-api-access-gx5ts") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "kube-api-access-gx5ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.470095 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.474957 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.476244 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.479274 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.483474 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.483992 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.484297 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.485025 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.489580 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4acf8a6b-4d95-435f-a697-49a803be1454" (UID: "4acf8a6b-4d95-435f-a697-49a803be1454"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554390 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554445 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554459 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554473 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554486 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554498 4846 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4acf8a6b-4d95-435f-a697-49a803be1454-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554512 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554524 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554537 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554550 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554562 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554574 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx5ts\" (UniqueName: \"kubernetes.io/projected/4acf8a6b-4d95-435f-a697-49a803be1454-kube-api-access-gx5ts\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554587 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.554598 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4acf8a6b-4d95-435f-a697-49a803be1454-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.720942 4846 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.744970 4846 generic.go:334] "Generic (PLEG): container finished" podID="4acf8a6b-4d95-435f-a697-49a803be1454" containerID="352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a" exitCode=0 Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.745048 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" event={"ID":"4acf8a6b-4d95-435f-a697-49a803be1454","Type":"ContainerDied","Data":"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a"} Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.745055 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.745093 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kgqq8" event={"ID":"4acf8a6b-4d95-435f-a697-49a803be1454","Type":"ContainerDied","Data":"3487b6391fa123c3205072262dd4a0c7aae12a82cf3715d7d6d489d862550368"} Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.745124 4846 scope.go:117] "RemoveContainer" containerID="352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.767230 4846 scope.go:117] "RemoveContainer" containerID="352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a" Mar 20 11:02:51 crc kubenswrapper[4846]: E0320 11:02:51.767727 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a\": container with ID starting with 352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a not found: ID does not exist" containerID="352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.767768 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a"} err="failed to get container status \"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a\": rpc error: code = NotFound desc = could not find container \"352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a\": container with ID starting with 352907e89215ba17889ecc2ac716b3b47dcf4a435c4442481d3c9e776953274a not found: ID does not exist" Mar 20 11:02:51 crc kubenswrapper[4846]: I0320 11:02:51.851377 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feb511e6-7df0-48d2-998b-76d26e4e0361" Mar 20 11:02:52 crc kubenswrapper[4846]: E0320 11:02:52.053141 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Mar 20 11:02:52 crc kubenswrapper[4846]: E0320 11:02:52.231542 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Mar 20 11:02:52 crc kubenswrapper[4846]: E0320 11:02:52.338403 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Mar 20 11:02:52 crc kubenswrapper[4846]: I0320 11:02:52.756518 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:52 crc kubenswrapper[4846]: I0320 11:02:52.757140 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:52 crc kubenswrapper[4846]: I0320 11:02:52.761807 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feb511e6-7df0-48d2-998b-76d26e4e0361" Mar 20 11:02:52 crc kubenswrapper[4846]: I0320 11:02:52.762793 4846 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://d1fd3c1b5b2302f4c7fb83f9df00530bc8935632b8699e0841b0119f86d64cba" Mar 20 11:02:52 crc kubenswrapper[4846]: I0320 11:02:52.763072 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:02:53 crc kubenswrapper[4846]: I0320 11:02:53.016762 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 11:02:53 crc kubenswrapper[4846]: I0320 11:02:53.763084 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:53 crc kubenswrapper[4846]: I0320 11:02:53.763133 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="447c448e-9ca0-45fa-b1d4-4bea6c3ef5fa" Mar 20 11:02:53 crc kubenswrapper[4846]: I0320 11:02:53.766749 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="feb511e6-7df0-48d2-998b-76d26e4e0361" Mar 20 11:03:01 crc kubenswrapper[4846]: I0320 11:03:01.321124 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 11:03:01 crc kubenswrapper[4846]: I0320 11:03:01.610419 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 11:03:01 crc kubenswrapper[4846]: I0320 11:03:01.723642 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 11:03:01 crc kubenswrapper[4846]: I0320 11:03:01.756546 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 11:03:01 crc kubenswrapper[4846]: I0320 11:03:01.916677 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.083201 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.246393 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.318351 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.583473 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.746182 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 11:03:02 crc kubenswrapper[4846]: I0320 11:03:02.999265 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.008540 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.512372 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.736596 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.862636 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.952354 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.996414 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 11:03:03 crc kubenswrapper[4846]: I0320 11:03:03.996557 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 11:03:04 crc kubenswrapper[4846]: I0320 11:03:04.458254 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 11:03:04 crc kubenswrapper[4846]: I0320 11:03:04.473956 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 11:03:04 crc kubenswrapper[4846]: I0320 11:03:04.633968 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 11:03:04 crc kubenswrapper[4846]: I0320 11:03:04.930401 4846 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.128760 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.168673 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.206941 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.216098 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.286405 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.317291 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.396347 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.401365 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.424742 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.430571 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.486071 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.489256 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.553836 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.559023 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.590309 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.656360 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.926330 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.959677 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.959766 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.959790 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.966261 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 11:03:05 crc kubenswrapper[4846]: I0320 11:03:05.969009 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.040377 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.070700 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.088355 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.158378 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.253079 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.317551 4846 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.343860 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.393747 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.400832 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.416845 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.417021 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.491376 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.551102 4846 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.579164 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.598841 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.707502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.731326 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.744715 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.744761 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.787599 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.851450 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.877854 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.891627 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.901238 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.955597 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.990983 4846 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.997328 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-kgqq8"] Mar 20 11:03:06 crc kubenswrapper[4846]: I0320 11:03:06.997440 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.002323 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.015284 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.015259433 podStartE2EDuration="16.015259433s" podCreationTimestamp="2026-03-20 11:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:03:07.014786971 +0000 UTC m=+358.381195207" watchObservedRunningTime="2026-03-20 11:03:07.015259433 +0000 UTC m=+358.381667669" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.022019 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.043876 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.091588 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.107260 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.120982 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.208389 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.283119 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.333192 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" path="/var/lib/kubelet/pods/4acf8a6b-4d95-435f-a697-49a803be1454/volumes" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.367927 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.400355 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.405394 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.430866 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.443569 4846 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.498452 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.564090 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.697153 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.737259 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.854576 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.854791 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.895093 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 11:03:07 crc kubenswrapper[4846]: I0320 11:03:07.927150 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.025843 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.053248 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.082749 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.088355 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.191152 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.304697 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.328049 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.350393 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.485649 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.488054 4846 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.488847 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.577661 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.638414 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.660502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.766976 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.834768 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.937570 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.939659 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 11:03:08 crc kubenswrapper[4846]: I0320 11:03:08.961929 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.069263 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.142726 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.291842 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.297916 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.505011 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.623392 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.639864 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.796987 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.827026 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.879295 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 11:03:09 crc kubenswrapper[4846]: I0320 11:03:09.987471 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.001133 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.098911 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.101771 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.104825 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.155354 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.212572 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.353698 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.372225 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.380220 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.388259 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.429767 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.466532 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.492949 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.572761 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.608461 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.672080 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.697046 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.734863 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.765807 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.788680 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.884471 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.891509 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 11:03:10 crc kubenswrapper[4846]: I0320 11:03:10.912522 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.024232 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.051198 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.127255 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.148768 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.154201 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.263961 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.293357 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.357687 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7cd676f696-6khrz"] Mar 20 11:03:11 crc kubenswrapper[4846]: E0320 11:03:11.357953 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" containerName="oauth-openshift" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.357969 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" containerName="oauth-openshift" Mar 20 11:03:11 crc kubenswrapper[4846]: E0320 11:03:11.357996 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" containerName="installer" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.358002 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" containerName="installer" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.358118 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa37f4d-5b65-45f4-ae84-a1f563508ff6" containerName="installer" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.358140 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4acf8a6b-4d95-435f-a697-49a803be1454" containerName="oauth-openshift" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.358604 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.363650 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.364012 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.364227 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.364404 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.364543 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.364953 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.365193 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.365194 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.365408 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.365942 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.369621 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.370053 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.372347 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.376428 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cd676f696-6khrz"] Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.433067 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.435747 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.436142 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.436272 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.438658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.454509 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.472013 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.494688 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.497511 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.550424 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-policies\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.550780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.550931 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551132 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hzcg\" (UniqueName: \"kubernetes.io/projected/d382e702-91da-49f1-a7a0-6c08eb401afa-kube-api-access-8hzcg\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551224 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551341 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551440 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-dir\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551615 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551716 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-session\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.551910 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.552099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.633305 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.649880 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hzcg\" (UniqueName: \"kubernetes.io/projected/d382e702-91da-49f1-a7a0-6c08eb401afa-kube-api-access-8hzcg\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653282 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-dir\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653312 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-session\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653423 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653480 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-policies\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653500 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.653531 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.654716 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-dir\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.655483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-audit-policies\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.656752 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-service-ca\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.657451 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.657960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.661189 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.661629 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-login\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.662571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-router-certs\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.671787 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hzcg\" (UniqueName: \"kubernetes.io/projected/d382e702-91da-49f1-a7a0-6c08eb401afa-kube-api-access-8hzcg\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.675038 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.677481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-template-error\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.681240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.683952 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.685943 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.691418 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d382e702-91da-49f1-a7a0-6c08eb401afa-v4-0-config-system-session\") pod \"oauth-openshift-7cd676f696-6khrz\" (UID: \"d382e702-91da-49f1-a7a0-6c08eb401afa\") " pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.732587 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.764346 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.872084 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 11:03:11 crc kubenswrapper[4846]: I0320 11:03:11.977362 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.059522 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.089814 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.140863 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.212774 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.262856 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.355461 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.389520 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.436042 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7cd676f696-6khrz"] Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.584313 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.647219 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.651082 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.670586 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.726413 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.763678 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.767467 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.851149 4846 scope.go:117] "RemoveContainer" containerID="2136c8ca988a73dff4252311b029fd0ce369bf84e05cd1e2baf9777f899b6ffb" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.872334 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.891191 4846 scope.go:117] "RemoveContainer" containerID="105b1818c5be467a837bf6dee188bf6bb08b8898525f39f09118bb9efef5ae5e" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.908558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" event={"ID":"d382e702-91da-49f1-a7a0-6c08eb401afa","Type":"ContainerStarted","Data":"49c4a23db55596ffdb33c22728ed0f0c2e766b139c71370342a6310b8fc80a6f"} Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.908629 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" event={"ID":"d382e702-91da-49f1-a7a0-6c08eb401afa","Type":"ContainerStarted","Data":"f3ac47233598ca0fb31f178b94f0b7f45b0de164243c1279ac841a749861df3c"} Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.909056 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.912174 4846 scope.go:117] "RemoveContainer" containerID="fe871630b134abc417ba03b3a886b070fa49da1740556be7aefaf8392f24f846" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.939251 4846 scope.go:117] "RemoveContainer" containerID="f9ed47ddade22d2b80075daffd2830cdb2a70bf76d9dade97efcd02026d5342d" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.941398 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" podStartSLOduration=47.941378537 podStartE2EDuration="47.941378537s" podCreationTimestamp="2026-03-20 11:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:03:12.936464146 +0000 UTC m=+364.302872382" watchObservedRunningTime="2026-03-20 11:03:12.941378537 +0000 UTC m=+364.307786773" Mar 20 11:03:12 crc kubenswrapper[4846]: I0320 11:03:12.959076 4846 scope.go:117] "RemoveContainer" containerID="93eeb627262d36e84c568e4ab44401a305cafdefe14fd56979c0e9030b815b71" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.149884 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.156041 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.167835 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7cd676f696-6khrz" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.175002 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.243231 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.253160 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.287535 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.293937 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.375147 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.405610 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.486273 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.501147 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.527162 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.548216 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.575422 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.715833 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.813738 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.842382 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.849019 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 11:03:13 crc kubenswrapper[4846]: I0320 11:03:13.850652 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.063428 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.171673 4846 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.173555 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c" gracePeriod=5 Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.267748 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.365889 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.397351 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.504839 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.603739 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.706850 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.896576 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 11:03:14 crc kubenswrapper[4846]: I0320 11:03:14.999720 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.023469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.164128 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.346049 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.624525 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.640938 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.700394 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 11:03:15 crc kubenswrapper[4846]: I0320 11:03:15.985076 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.105821 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.163780 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.164171 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.225452 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.291347 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.443351 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.594404 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.619961 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.637376 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.650140 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.673931 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.689176 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.696047 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.849261 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.860064 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.948453 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 11:03:16 crc kubenswrapper[4846]: I0320 11:03:16.987434 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 11:03:17 crc kubenswrapper[4846]: I0320 11:03:17.690341 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 11:03:17 crc kubenswrapper[4846]: I0320 11:03:17.840451 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 11:03:18 crc kubenswrapper[4846]: I0320 11:03:18.273416 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.806745 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.807066 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.956051 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.956120 4846 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c" exitCode=137 Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.956184 4846 scope.go:117] "RemoveContainer" containerID="9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.956451 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.974964 4846 scope.go:117] "RemoveContainer" containerID="9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c" Mar 20 11:03:19 crc kubenswrapper[4846]: E0320 11:03:19.975491 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c\": container with ID starting with 9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c not found: ID does not exist" containerID="9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.975602 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c"} err="failed to get container status \"9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c\": rpc error: code = NotFound desc = could not find container \"9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c\": container with ID starting with 9cea50683c45fa3b6240b05ce8f023f1fc1432756d87da6a6cbf06110f323f2c not found: ID does not exist" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978357 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978461 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978648 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978684 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.978827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.979044 4846 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.979076 4846 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.979092 4846 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.979106 4846 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 20 11:03:19 crc kubenswrapper[4846]: I0320 11:03:19.987080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:03:20 crc kubenswrapper[4846]: I0320 11:03:20.080621 4846 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 11:03:21 crc kubenswrapper[4846]: I0320 11:03:21.332891 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 20 11:03:29 crc kubenswrapper[4846]: I0320 11:03:29.888021 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 11:03:33 crc kubenswrapper[4846]: I0320 11:03:33.219362 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 11:03:33 crc kubenswrapper[4846]: I0320 11:03:33.269942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerDied","Data":"7323dedc584f0a1300da63771d0bd4af4e018a00c6e7ae60fbb8d1b55efa548d"} Mar 20 11:03:33 crc kubenswrapper[4846]: I0320 11:03:33.269883 4846 generic.go:334] "Generic (PLEG): container finished" podID="9590f0a5-0754-484b-9f20-89e6155722d6" containerID="7323dedc584f0a1300da63771d0bd4af4e018a00c6e7ae60fbb8d1b55efa548d" exitCode=0 Mar 20 11:03:33 crc kubenswrapper[4846]: I0320 11:03:33.270946 4846 scope.go:117] "RemoveContainer" containerID="7323dedc584f0a1300da63771d0bd4af4e018a00c6e7ae60fbb8d1b55efa548d" Mar 20 11:03:34 crc kubenswrapper[4846]: I0320 11:03:34.278171 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerStarted","Data":"7722d1bc9a0f5a66664d96d60cd3cc0f8782115391203bbf8bffd025c897c5a6"} Mar 20 11:03:34 crc kubenswrapper[4846]: I0320 11:03:34.278532 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:03:34 crc kubenswrapper[4846]: I0320 11:03:34.281932 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:03:53 crc kubenswrapper[4846]: I0320 11:03:53.426569 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 11:03:57 crc kubenswrapper[4846]: I0320 11:03:57.575475 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.176130 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566744-n8dcz"] Mar 20 11:04:00 crc kubenswrapper[4846]: E0320 11:04:00.176954 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.176987 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.177263 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.178198 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.184238 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.184240 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566744-n8dcz"] Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.184748 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.185166 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.193359 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvshc\" (UniqueName: \"kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc\") pod \"auto-csr-approver-29566744-n8dcz\" (UID: \"cf95362a-3893-4688-8c22-7092d445382a\") " pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.294687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvshc\" (UniqueName: \"kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc\") pod \"auto-csr-approver-29566744-n8dcz\" (UID: \"cf95362a-3893-4688-8c22-7092d445382a\") " pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.333813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvshc\" (UniqueName: \"kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc\") pod \"auto-csr-approver-29566744-n8dcz\" (UID: \"cf95362a-3893-4688-8c22-7092d445382a\") " pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.506807 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:00 crc kubenswrapper[4846]: I0320 11:04:00.938292 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566744-n8dcz"] Mar 20 11:04:01 crc kubenswrapper[4846]: I0320 11:04:01.466065 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" event={"ID":"cf95362a-3893-4688-8c22-7092d445382a","Type":"ContainerStarted","Data":"b98188ab313dbe2ddd512b8b9b62acf1bb4737af55e540be4a6af0e2417fe294"} Mar 20 11:04:02 crc kubenswrapper[4846]: I0320 11:04:02.473393 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" event={"ID":"cf95362a-3893-4688-8c22-7092d445382a","Type":"ContainerStarted","Data":"347a272e87ec1fb8bd4cea9f378d7adf936795fbc3ab3bae6eaf68fa9afa595a"} Mar 20 11:04:02 crc kubenswrapper[4846]: I0320 11:04:02.487544 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" podStartSLOduration=1.419847361 podStartE2EDuration="2.487517066s" podCreationTimestamp="2026-03-20 11:04:00 +0000 UTC" firstStartedPulling="2026-03-20 11:04:00.948976164 +0000 UTC m=+412.315384400" lastFinishedPulling="2026-03-20 11:04:02.016645869 +0000 UTC m=+413.383054105" observedRunningTime="2026-03-20 11:04:02.485018119 +0000 UTC m=+413.851426365" watchObservedRunningTime="2026-03-20 11:04:02.487517066 +0000 UTC m=+413.853925302" Mar 20 11:04:03 crc kubenswrapper[4846]: I0320 11:04:03.482565 4846 generic.go:334] "Generic (PLEG): container finished" podID="cf95362a-3893-4688-8c22-7092d445382a" containerID="347a272e87ec1fb8bd4cea9f378d7adf936795fbc3ab3bae6eaf68fa9afa595a" exitCode=0 Mar 20 11:04:03 crc kubenswrapper[4846]: I0320 11:04:03.482632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" event={"ID":"cf95362a-3893-4688-8c22-7092d445382a","Type":"ContainerDied","Data":"347a272e87ec1fb8bd4cea9f378d7adf936795fbc3ab3bae6eaf68fa9afa595a"} Mar 20 11:04:04 crc kubenswrapper[4846]: I0320 11:04:04.855433 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:04 crc kubenswrapper[4846]: I0320 11:04:04.861496 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvshc\" (UniqueName: \"kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc\") pod \"cf95362a-3893-4688-8c22-7092d445382a\" (UID: \"cf95362a-3893-4688-8c22-7092d445382a\") " Mar 20 11:04:04 crc kubenswrapper[4846]: I0320 11:04:04.868409 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc" (OuterVolumeSpecName: "kube-api-access-cvshc") pod "cf95362a-3893-4688-8c22-7092d445382a" (UID: "cf95362a-3893-4688-8c22-7092d445382a"). InnerVolumeSpecName "kube-api-access-cvshc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:04:04 crc kubenswrapper[4846]: I0320 11:04:04.963883 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvshc\" (UniqueName: \"kubernetes.io/projected/cf95362a-3893-4688-8c22-7092d445382a-kube-api-access-cvshc\") on node \"crc\" DevicePath \"\"" Mar 20 11:04:05 crc kubenswrapper[4846]: I0320 11:04:05.495850 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" event={"ID":"cf95362a-3893-4688-8c22-7092d445382a","Type":"ContainerDied","Data":"b98188ab313dbe2ddd512b8b9b62acf1bb4737af55e540be4a6af0e2417fe294"} Mar 20 11:04:05 crc kubenswrapper[4846]: I0320 11:04:05.495944 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b98188ab313dbe2ddd512b8b9b62acf1bb4737af55e540be4a6af0e2417fe294" Mar 20 11:04:05 crc kubenswrapper[4846]: I0320 11:04:05.495966 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566744-n8dcz" Mar 20 11:04:39 crc kubenswrapper[4846]: I0320 11:04:39.677416 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:04:39 crc kubenswrapper[4846]: I0320 11:04:39.678192 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:04:40 crc kubenswrapper[4846]: I0320 11:04:40.971068 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bxtnx"] Mar 20 11:04:40 crc kubenswrapper[4846]: E0320 11:04:40.971403 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf95362a-3893-4688-8c22-7092d445382a" containerName="oc" Mar 20 11:04:40 crc kubenswrapper[4846]: I0320 11:04:40.971419 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf95362a-3893-4688-8c22-7092d445382a" containerName="oc" Mar 20 11:04:40 crc kubenswrapper[4846]: I0320 11:04:40.971576 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf95362a-3893-4688-8c22-7092d445382a" containerName="oc" Mar 20 11:04:40 crc kubenswrapper[4846]: I0320 11:04:40.972186 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:40 crc kubenswrapper[4846]: I0320 11:04:40.993438 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bxtnx"] Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111392 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-tls\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111447 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111492 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-bound-sa-token\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111511 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-certificates\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111840 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-trusted-ca\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.111997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxwcz\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-kube-api-access-pxwcz\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.133357 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.213324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-trusted-ca\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.213388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxwcz\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-kube-api-access-pxwcz\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.213420 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-tls\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.213465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.214415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.214580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-bound-sa-token\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.214751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-certificates\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.215667 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.215688 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-trusted-ca\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.216451 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-certificates\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.222209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.225521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-registry-tls\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.232079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxwcz\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-kube-api-access-pxwcz\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.236718 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27-bound-sa-token\") pod \"image-registry-66df7c8f76-bxtnx\" (UID: \"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27\") " pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.291505 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.551169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bxtnx"] Mar 20 11:04:41 crc kubenswrapper[4846]: I0320 11:04:41.717997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" event={"ID":"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27","Type":"ContainerStarted","Data":"717d17ff6493acb9d99abb59e3851dbab692367d4d1fdee4a39d669ab1dedd78"} Mar 20 11:04:42 crc kubenswrapper[4846]: I0320 11:04:42.727097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" event={"ID":"e2d851b2-a2bf-4f7d-9881-37b7fb3d1d27","Type":"ContainerStarted","Data":"10698386234f8eb8d4249f7773084d819087769a4388ac6c9aa771e45d242042"} Mar 20 11:04:42 crc kubenswrapper[4846]: I0320 11:04:42.727322 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:04:42 crc kubenswrapper[4846]: I0320 11:04:42.761163 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" podStartSLOduration=2.7611320360000002 podStartE2EDuration="2.761132036s" podCreationTimestamp="2026-03-20 11:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:04:42.754823762 +0000 UTC m=+454.121232068" watchObservedRunningTime="2026-03-20 11:04:42.761132036 +0000 UTC m=+454.127540282" Mar 20 11:05:01 crc kubenswrapper[4846]: I0320 11:05:01.300323 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bxtnx" Mar 20 11:05:01 crc kubenswrapper[4846]: I0320 11:05:01.383058 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:05:09 crc kubenswrapper[4846]: I0320 11:05:09.677994 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:05:09 crc kubenswrapper[4846]: I0320 11:05:09.678606 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.422506 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.423598 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tt9fr" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="registry-server" containerID="cri-o://0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198" gracePeriod=30 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.431886 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.432434 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-56jhc" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="registry-server" containerID="cri-o://a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c" gracePeriod=30 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.448858 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.457300 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.457562 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" containerID="cri-o://7722d1bc9a0f5a66664d96d60cd3cc0f8782115391203bbf8bffd025c897c5a6" gracePeriod=30 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.458312 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6vwj9" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="registry-server" containerID="cri-o://654e75bc0015d51a0a4f8cd4fc259f6c082039319f43ee1f2e05d520b54a8d47" gracePeriod=30 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.472178 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.472819 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kh2d5" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="registry-server" containerID="cri-o://7694dc1764c7328d2c31062299e1d6cd0fe863b3ba7083df719e4d4d21900533" gracePeriod=30 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.482919 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z5nmm"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.483785 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.530185 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z5nmm"] Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.624742 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsm2m\" (UniqueName: \"kubernetes.io/projected/46f25710-1380-49bc-9a98-84a7b4f1f455-kube-api-access-lsm2m\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.625077 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.625124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.726403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsm2m\" (UniqueName: \"kubernetes.io/projected/46f25710-1380-49bc-9a98-84a7b4f1f455-kube-api-access-lsm2m\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.726480 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.726546 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.728240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.736734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46f25710-1380-49bc-9a98-84a7b4f1f455-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.746046 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsm2m\" (UniqueName: \"kubernetes.io/projected/46f25710-1380-49bc-9a98-84a7b4f1f455-kube-api-access-lsm2m\") pod \"marketplace-operator-79b997595-z5nmm\" (UID: \"46f25710-1380-49bc-9a98-84a7b4f1f455\") " pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.888487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.897538 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.902584 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.925979 4846 generic.go:334] "Generic (PLEG): container finished" podID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerID="0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198" exitCode=0 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.926073 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerDied","Data":"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.926078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt9fr" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.926114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt9fr" event={"ID":"d6291c67-3539-45a8-bbb5-312be6bebcd4","Type":"ContainerDied","Data":"524dbe9258cd5b4f00bb342c6667ea970055de8fab93d7b0473e9ab8dfac6a7b"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.926136 4846 scope.go:117] "RemoveContainer" containerID="0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.928560 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerID="654e75bc0015d51a0a4f8cd4fc259f6c082039319f43ee1f2e05d520b54a8d47" exitCode=0 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.928632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerDied","Data":"654e75bc0015d51a0a4f8cd4fc259f6c082039319f43ee1f2e05d520b54a8d47"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.928671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vwj9" event={"ID":"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7","Type":"ContainerDied","Data":"1e89c210582146a3985d6ff388bba58b9bc4a545217d0730ee1cef8c249e1e37"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.928687 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e89c210582146a3985d6ff388bba58b9bc4a545217d0730ee1cef8c249e1e37" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.931715 4846 generic.go:334] "Generic (PLEG): container finished" podID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerID="a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c" exitCode=0 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.931764 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerDied","Data":"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.931789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56jhc" event={"ID":"2b17dfdf-1984-4cde-958e-021958ad8e0c","Type":"ContainerDied","Data":"faa3a5a937d1a11e52535d180ca9b905119494c913197e43b6867d65abeff89f"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.931872 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56jhc" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.933639 4846 generic.go:334] "Generic (PLEG): container finished" podID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerID="7694dc1764c7328d2c31062299e1d6cd0fe863b3ba7083df719e4d4d21900533" exitCode=0 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.933692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerDied","Data":"7694dc1764c7328d2c31062299e1d6cd0fe863b3ba7083df719e4d4d21900533"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.933712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kh2d5" event={"ID":"de1fb688-1d64-4289-9d68-faf88d479fa1","Type":"ContainerDied","Data":"a189629dffd267da69a8916dcc27fe031e6418fbaa757b896e8df4a12c75e77e"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.933725 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a189629dffd267da69a8916dcc27fe031e6418fbaa757b896e8df4a12c75e77e" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.940213 4846 generic.go:334] "Generic (PLEG): container finished" podID="9590f0a5-0754-484b-9f20-89e6155722d6" containerID="7722d1bc9a0f5a66664d96d60cd3cc0f8782115391203bbf8bffd025c897c5a6" exitCode=0 Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.940271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerDied","Data":"7722d1bc9a0f5a66664d96d60cd3cc0f8782115391203bbf8bffd025c897c5a6"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.940325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" event={"ID":"9590f0a5-0754-484b-9f20-89e6155722d6","Type":"ContainerDied","Data":"154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd"} Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.940340 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="154f933154a7c513fd0ccdedbd6c017aa513fb8c5268e670aed567d5955131cd" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.956381 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.975022 4846 scope.go:117] "RemoveContainer" containerID="7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.986172 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:05:10 crc kubenswrapper[4846]: I0320 11:05:10.999247 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.007860 4846 scope.go:117] "RemoveContainer" containerID="27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.030871 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities\") pod \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.030939 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj6gz\" (UniqueName: \"kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz\") pod \"d6291c67-3539-45a8-bbb5-312be6bebcd4\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.030961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content\") pod \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.030986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content\") pod \"2b17dfdf-1984-4cde-958e-021958ad8e0c\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.031028 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities\") pod \"d6291c67-3539-45a8-bbb5-312be6bebcd4\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.031096 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities\") pod \"2b17dfdf-1984-4cde-958e-021958ad8e0c\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.031214 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s4jw\" (UniqueName: \"kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw\") pod \"2b17dfdf-1984-4cde-958e-021958ad8e0c\" (UID: \"2b17dfdf-1984-4cde-958e-021958ad8e0c\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.031275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p45cb\" (UniqueName: \"kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb\") pod \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\" (UID: \"bb0fabe2-dd2f-451f-8f04-68feddb0fdd7\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.031366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content\") pod \"d6291c67-3539-45a8-bbb5-312be6bebcd4\" (UID: \"d6291c67-3539-45a8-bbb5-312be6bebcd4\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.032257 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities" (OuterVolumeSpecName: "utilities") pod "d6291c67-3539-45a8-bbb5-312be6bebcd4" (UID: "d6291c67-3539-45a8-bbb5-312be6bebcd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.032919 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities" (OuterVolumeSpecName: "utilities") pod "2b17dfdf-1984-4cde-958e-021958ad8e0c" (UID: "2b17dfdf-1984-4cde-958e-021958ad8e0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.040771 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb" (OuterVolumeSpecName: "kube-api-access-p45cb") pod "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" (UID: "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7"). InnerVolumeSpecName "kube-api-access-p45cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.042188 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw" (OuterVolumeSpecName: "kube-api-access-4s4jw") pod "2b17dfdf-1984-4cde-958e-021958ad8e0c" (UID: "2b17dfdf-1984-4cde-958e-021958ad8e0c"). InnerVolumeSpecName "kube-api-access-4s4jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.043932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz" (OuterVolumeSpecName: "kube-api-access-vj6gz") pod "d6291c67-3539-45a8-bbb5-312be6bebcd4" (UID: "d6291c67-3539-45a8-bbb5-312be6bebcd4"). InnerVolumeSpecName "kube-api-access-vj6gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.044452 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities" (OuterVolumeSpecName: "utilities") pod "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" (UID: "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.066185 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" (UID: "bb0fabe2-dd2f-451f-8f04-68feddb0fdd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.084454 4846 scope.go:117] "RemoveContainer" containerID="0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.084930 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198\": container with ID starting with 0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198 not found: ID does not exist" containerID="0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.084973 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198"} err="failed to get container status \"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198\": rpc error: code = NotFound desc = could not find container \"0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198\": container with ID starting with 0c87fd4e9e61fb812ecce67655abce2cc737b6cbd47800c137c1c042ad5ad198 not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.085002 4846 scope.go:117] "RemoveContainer" containerID="7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.085841 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1\": container with ID starting with 7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1 not found: ID does not exist" containerID="7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.085870 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1"} err="failed to get container status \"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1\": rpc error: code = NotFound desc = could not find container \"7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1\": container with ID starting with 7f63f7d7eb454f7ce7a33aa9d9a622b43574b8eefe6d5d543de515f32e8e15b1 not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.085887 4846 scope.go:117] "RemoveContainer" containerID="27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.086063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b17dfdf-1984-4cde-958e-021958ad8e0c" (UID: "2b17dfdf-1984-4cde-958e-021958ad8e0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.086440 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4\": container with ID starting with 27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4 not found: ID does not exist" containerID="27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.086469 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4"} err="failed to get container status \"27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4\": rpc error: code = NotFound desc = could not find container \"27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4\": container with ID starting with 27fc94e6a35b1b90cb3c43ce88d38b47ac98ad4838446e47a82d89f8edda40d4 not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.086487 4846 scope.go:117] "RemoveContainer" containerID="a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.091808 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6291c67-3539-45a8-bbb5-312be6bebcd4" (UID: "d6291c67-3539-45a8-bbb5-312be6bebcd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.109279 4846 scope.go:117] "RemoveContainer" containerID="3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.122004 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z5nmm"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.127207 4846 scope.go:117] "RemoveContainer" containerID="d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.132597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities\") pod \"de1fb688-1d64-4289-9d68-faf88d479fa1\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.133117 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjvlg\" (UniqueName: \"kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg\") pod \"de1fb688-1d64-4289-9d68-faf88d479fa1\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.133182 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca\") pod \"9590f0a5-0754-484b-9f20-89e6155722d6\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.133843 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics\") pod \"9590f0a5-0754-484b-9f20-89e6155722d6\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.133747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9590f0a5-0754-484b-9f20-89e6155722d6" (UID: "9590f0a5-0754-484b-9f20-89e6155722d6"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.134383 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smlwg\" (UniqueName: \"kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg\") pod \"9590f0a5-0754-484b-9f20-89e6155722d6\" (UID: \"9590f0a5-0754-484b-9f20-89e6155722d6\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.134421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content\") pod \"de1fb688-1d64-4289-9d68-faf88d479fa1\" (UID: \"de1fb688-1d64-4289-9d68-faf88d479fa1\") " Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.134370 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities" (OuterVolumeSpecName: "utilities") pod "de1fb688-1d64-4289-9d68-faf88d479fa1" (UID: "de1fb688-1d64-4289-9d68-faf88d479fa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.136347 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg" (OuterVolumeSpecName: "kube-api-access-xjvlg") pod "de1fb688-1d64-4289-9d68-faf88d479fa1" (UID: "de1fb688-1d64-4289-9d68-faf88d479fa1"). InnerVolumeSpecName "kube-api-access-xjvlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.136859 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9590f0a5-0754-484b-9f20-89e6155722d6" (UID: "9590f0a5-0754-484b-9f20-89e6155722d6"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140393 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140444 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140460 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjvlg\" (UniqueName: \"kubernetes.io/projected/de1fb688-1d64-4289-9d68-faf88d479fa1-kube-api-access-xjvlg\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140471 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140482 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9590f0a5-0754-484b-9f20-89e6155722d6-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140497 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140509 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140520 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj6gz\" (UniqueName: \"kubernetes.io/projected/d6291c67-3539-45a8-bbb5-312be6bebcd4-kube-api-access-vj6gz\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140537 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140547 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6291c67-3539-45a8-bbb5-312be6bebcd4-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140558 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b17dfdf-1984-4cde-958e-021958ad8e0c-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140568 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s4jw\" (UniqueName: \"kubernetes.io/projected/2b17dfdf-1984-4cde-958e-021958ad8e0c-kube-api-access-4s4jw\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.140582 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p45cb\" (UniqueName: \"kubernetes.io/projected/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7-kube-api-access-p45cb\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.146534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg" (OuterVolumeSpecName: "kube-api-access-smlwg") pod "9590f0a5-0754-484b-9f20-89e6155722d6" (UID: "9590f0a5-0754-484b-9f20-89e6155722d6"). InnerVolumeSpecName "kube-api-access-smlwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.150733 4846 scope.go:117] "RemoveContainer" containerID="a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.157379 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c\": container with ID starting with a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c not found: ID does not exist" containerID="a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.157441 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c"} err="failed to get container status \"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c\": rpc error: code = NotFound desc = could not find container \"a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c\": container with ID starting with a9ba3a74557128e5e207e3d5e869d503fc56e4f6ce704b857442645341b2449c not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.157495 4846 scope.go:117] "RemoveContainer" containerID="3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.157974 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210\": container with ID starting with 3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210 not found: ID does not exist" containerID="3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.158012 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210"} err="failed to get container status \"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210\": rpc error: code = NotFound desc = could not find container \"3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210\": container with ID starting with 3d4d9cc0487f802a3459a36cd45f97dc1c545745190f11ea4416d5b6715c3210 not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.158034 4846 scope.go:117] "RemoveContainer" containerID="d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9" Mar 20 11:05:11 crc kubenswrapper[4846]: E0320 11:05:11.158655 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9\": container with ID starting with d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9 not found: ID does not exist" containerID="d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.158685 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9"} err="failed to get container status \"d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9\": rpc error: code = NotFound desc = could not find container \"d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9\": container with ID starting with d729fd25c21647c0904b1b4b145c4ef509a444dc73668c91266d82b3029bf8c9 not found: ID does not exist" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.158734 4846 scope.go:117] "RemoveContainer" containerID="7323dedc584f0a1300da63771d0bd4af4e018a00c6e7ae60fbb8d1b55efa548d" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.241429 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smlwg\" (UniqueName: \"kubernetes.io/projected/9590f0a5-0754-484b-9f20-89e6155722d6-kube-api-access-smlwg\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.272059 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.274217 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tt9fr"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.278837 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.283611 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-56jhc"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.292735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de1fb688-1d64-4289-9d68-faf88d479fa1" (UID: "de1fb688-1d64-4289-9d68-faf88d479fa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.339800 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" path="/var/lib/kubelet/pods/2b17dfdf-1984-4cde-958e-021958ad8e0c/volumes" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.340568 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" path="/var/lib/kubelet/pods/d6291c67-3539-45a8-bbb5-312be6bebcd4/volumes" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.342415 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1fb688-1d64-4289-9d68-faf88d479fa1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.948267 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5phhj" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.952912 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" event={"ID":"46f25710-1380-49bc-9a98-84a7b4f1f455","Type":"ContainerStarted","Data":"def6afee9ec2df158c151e6a6d1c6de063f1347057e2b9e87d4df0f7f695ca9e"} Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.952959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" event={"ID":"46f25710-1380-49bc-9a98-84a7b4f1f455","Type":"ContainerStarted","Data":"1cc501195a3cae840e1c83596bb11a7ecae93b3f34e2e6d2eacb1b4e98c83148"} Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.953004 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vwj9" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.953057 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kh2d5" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.955940 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.959695 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.983597 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.988365 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kh2d5"] Mar 20 11:05:11 crc kubenswrapper[4846]: I0320 11:05:11.995064 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z5nmm" podStartSLOduration=1.995046943 podStartE2EDuration="1.995046943s" podCreationTimestamp="2026-03-20 11:05:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:05:11.994679275 +0000 UTC m=+483.361087521" watchObservedRunningTime="2026-03-20 11:05:11.995046943 +0000 UTC m=+483.361455189" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.011134 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.016623 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5phhj"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.043653 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.050441 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vwj9"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.623796 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mkqk"] Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.624734 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.624805 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.624889 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.624999 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625070 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625133 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625209 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625280 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625361 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625428 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625503 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625568 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625645 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625725 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625805 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.625878 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.625979 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626050 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.626124 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626187 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="extract-utilities" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.626258 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626325 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.626391 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626480 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.626551 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626620 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: E0320 11:05:12.626687 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.626779 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="extract-content" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627000 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627085 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6291c67-3539-45a8-bbb5-312be6bebcd4" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627157 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627222 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b17dfdf-1984-4cde-958e-021958ad8e0c" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627288 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" containerName="marketplace-operator" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.627368 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" containerName="registry-server" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.628256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.639176 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mkqk"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.639686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.770010 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-catalog-content\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.770118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-utilities\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.770655 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldm2c\" (UniqueName: \"kubernetes.io/projected/4df61cc9-7ff8-47a8-9800-5f27801924f3-kube-api-access-ldm2c\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.852586 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.854625 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.874033 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-catalog-content\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.874170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-utilities\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.874206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldm2c\" (UniqueName: \"kubernetes.io/projected/4df61cc9-7ff8-47a8-9800-5f27801924f3-kube-api-access-ldm2c\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.874068 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.875162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-catalog-content\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.875204 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4df61cc9-7ff8-47a8-9800-5f27801924f3-utilities\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.878964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.935108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldm2c\" (UniqueName: \"kubernetes.io/projected/4df61cc9-7ff8-47a8-9800-5f27801924f3-kube-api-access-ldm2c\") pod \"certified-operators-7mkqk\" (UID: \"4df61cc9-7ff8-47a8-9800-5f27801924f3\") " pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.975334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4j7x\" (UniqueName: \"kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.975389 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.975452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:12 crc kubenswrapper[4846]: I0320 11:05:12.987984 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.076861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.077311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4j7x\" (UniqueName: \"kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.077342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.077804 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.078814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.103771 4846 scope.go:117] "RemoveContainer" containerID="c560d820c591ea55ba4c6122bae40f6e16ccd90bff8c1b4ea1ce37f242ee35c4" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.103865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4j7x\" (UniqueName: \"kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x\") pod \"community-operators-zb6fs\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.192311 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.196790 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mkqk"] Mar 20 11:05:13 crc kubenswrapper[4846]: W0320 11:05:13.206671 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4df61cc9_7ff8_47a8_9800_5f27801924f3.slice/crio-341dd023436a596c0e51c431eb442917e63cc5420c4ef0d63396f6755b29755d WatchSource:0}: Error finding container 341dd023436a596c0e51c431eb442917e63cc5420c4ef0d63396f6755b29755d: Status 404 returned error can't find the container with id 341dd023436a596c0e51c431eb442917e63cc5420c4ef0d63396f6755b29755d Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.333284 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9590f0a5-0754-484b-9f20-89e6155722d6" path="/var/lib/kubelet/pods/9590f0a5-0754-484b-9f20-89e6155722d6/volumes" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.333827 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb0fabe2-dd2f-451f-8f04-68feddb0fdd7" path="/var/lib/kubelet/pods/bb0fabe2-dd2f-451f-8f04-68feddb0fdd7/volumes" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.334507 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1fb688-1d64-4289-9d68-faf88d479fa1" path="/var/lib/kubelet/pods/de1fb688-1d64-4289-9d68-faf88d479fa1/volumes" Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.400790 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:05:13 crc kubenswrapper[4846]: W0320 11:05:13.444851 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6105d963_9e35_4d0b_897f_8b4488d105ba.slice/crio-a105c0b337bff13800c441a9c0442dbfef2d7470898656e43a5c71cb04108b6a WatchSource:0}: Error finding container a105c0b337bff13800c441a9c0442dbfef2d7470898656e43a5c71cb04108b6a: Status 404 returned error can't find the container with id a105c0b337bff13800c441a9c0442dbfef2d7470898656e43a5c71cb04108b6a Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.967210 4846 generic.go:334] "Generic (PLEG): container finished" podID="4df61cc9-7ff8-47a8-9800-5f27801924f3" containerID="da9639c4704530247368ed75e8fea099d4efe4b4f38cb9066b55cf552903b637" exitCode=0 Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.967257 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkqk" event={"ID":"4df61cc9-7ff8-47a8-9800-5f27801924f3","Type":"ContainerDied","Data":"da9639c4704530247368ed75e8fea099d4efe4b4f38cb9066b55cf552903b637"} Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.967307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkqk" event={"ID":"4df61cc9-7ff8-47a8-9800-5f27801924f3","Type":"ContainerStarted","Data":"341dd023436a596c0e51c431eb442917e63cc5420c4ef0d63396f6755b29755d"} Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.969131 4846 generic.go:334] "Generic (PLEG): container finished" podID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerID="4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca" exitCode=0 Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.969191 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerDied","Data":"4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca"} Mar 20 11:05:13 crc kubenswrapper[4846]: I0320 11:05:13.969273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerStarted","Data":"a105c0b337bff13800c441a9c0442dbfef2d7470898656e43a5c71cb04108b6a"} Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.030395 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nqpzh"] Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.031992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.034566 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.040465 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqpzh"] Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.111629 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-utilities\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.111683 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-catalog-content\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.111715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2xbd\" (UniqueName: \"kubernetes.io/projected/30a01aab-f965-4b19-97b5-a3bccd0aaff3-kube-api-access-q2xbd\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.215681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-utilities\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.215762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-catalog-content\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.215802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2xbd\" (UniqueName: \"kubernetes.io/projected/30a01aab-f965-4b19-97b5-a3bccd0aaff3-kube-api-access-q2xbd\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.216973 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-utilities\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.218318 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a01aab-f965-4b19-97b5-a3bccd0aaff3-catalog-content\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.232301 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bcsxf"] Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.233794 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.239807 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.242335 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcsxf"] Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.256047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2xbd\" (UniqueName: \"kubernetes.io/projected/30a01aab-f965-4b19-97b5-a3bccd0aaff3-kube-api-access-q2xbd\") pod \"redhat-marketplace-nqpzh\" (UID: \"30a01aab-f965-4b19-97b5-a3bccd0aaff3\") " pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.317065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-utilities\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.317150 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-catalog-content\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.317175 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6rzq\" (UniqueName: \"kubernetes.io/projected/70703261-65d2-40da-aa0d-948663787c75-kube-api-access-l6rzq\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.418638 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-utilities\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.418751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-catalog-content\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.418792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6rzq\" (UniqueName: \"kubernetes.io/projected/70703261-65d2-40da-aa0d-948663787c75-kube-api-access-l6rzq\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.420355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-catalog-content\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.422100 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70703261-65d2-40da-aa0d-948663787c75-utilities\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.443273 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6rzq\" (UniqueName: \"kubernetes.io/projected/70703261-65d2-40da-aa0d-948663787c75-kube-api-access-l6rzq\") pod \"redhat-operators-bcsxf\" (UID: \"70703261-65d2-40da-aa0d-948663787c75\") " pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.447697 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:15 crc kubenswrapper[4846]: I0320 11:05:15.619847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.798468 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcsxf"] Mar 20 11:05:16 crc kubenswrapper[4846]: W0320 11:05:15.808022 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70703261_65d2_40da_aa0d_948663787c75.slice/crio-1a3ea91170a427f98f6b8095bbb94146434d0073fcabae6dd5df420fa6318509 WatchSource:0}: Error finding container 1a3ea91170a427f98f6b8095bbb94146434d0073fcabae6dd5df420fa6318509: Status 404 returned error can't find the container with id 1a3ea91170a427f98f6b8095bbb94146434d0073fcabae6dd5df420fa6318509 Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.853432 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nqpzh"] Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.988639 4846 generic.go:334] "Generic (PLEG): container finished" podID="4df61cc9-7ff8-47a8-9800-5f27801924f3" containerID="266a168d4b16dbbc778ac43026abd593064efcc9aa7d270ffc32b59e39270963" exitCode=0 Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.988754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkqk" event={"ID":"4df61cc9-7ff8-47a8-9800-5f27801924f3","Type":"ContainerDied","Data":"266a168d4b16dbbc778ac43026abd593064efcc9aa7d270ffc32b59e39270963"} Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.994438 4846 generic.go:334] "Generic (PLEG): container finished" podID="70703261-65d2-40da-aa0d-948663787c75" containerID="78aeb8bffd9d7bde6742f12d4dfab1f6adb5f03d11b4818d3a557d785ebc6446" exitCode=0 Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.994618 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcsxf" event={"ID":"70703261-65d2-40da-aa0d-948663787c75","Type":"ContainerDied","Data":"78aeb8bffd9d7bde6742f12d4dfab1f6adb5f03d11b4818d3a557d785ebc6446"} Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.994652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcsxf" event={"ID":"70703261-65d2-40da-aa0d-948663787c75","Type":"ContainerStarted","Data":"1a3ea91170a427f98f6b8095bbb94146434d0073fcabae6dd5df420fa6318509"} Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:15.997657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqpzh" event={"ID":"30a01aab-f965-4b19-97b5-a3bccd0aaff3","Type":"ContainerStarted","Data":"3a3fdbf00c2ea9f510cb45f5fad2e65c0b29aa2e11add85e2f1c064e70c0665c"} Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:16.000838 4846 generic.go:334] "Generic (PLEG): container finished" podID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerID="4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241" exitCode=0 Mar 20 11:05:16 crc kubenswrapper[4846]: I0320 11:05:16.000913 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerDied","Data":"4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241"} Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.021414 4846 generic.go:334] "Generic (PLEG): container finished" podID="30a01aab-f965-4b19-97b5-a3bccd0aaff3" containerID="f07dac8a3a374f5df277c3cea162e4606c11426aa5e532c0ddb59c25a7d2ca0f" exitCode=0 Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.021547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqpzh" event={"ID":"30a01aab-f965-4b19-97b5-a3bccd0aaff3","Type":"ContainerDied","Data":"f07dac8a3a374f5df277c3cea162e4606c11426aa5e532c0ddb59c25a7d2ca0f"} Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.036594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mkqk" event={"ID":"4df61cc9-7ff8-47a8-9800-5f27801924f3","Type":"ContainerStarted","Data":"1c19ceee5ceefeaf37551bd3e3315140db9df0d5e218f721b19fa11431f892a5"} Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.039854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerStarted","Data":"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5"} Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.066168 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mkqk" podStartSLOduration=2.495805323 podStartE2EDuration="5.066137969s" podCreationTimestamp="2026-03-20 11:05:12 +0000 UTC" firstStartedPulling="2026-03-20 11:05:13.971051355 +0000 UTC m=+485.337459591" lastFinishedPulling="2026-03-20 11:05:16.541384001 +0000 UTC m=+487.907792237" observedRunningTime="2026-03-20 11:05:17.061496294 +0000 UTC m=+488.427904530" watchObservedRunningTime="2026-03-20 11:05:17.066137969 +0000 UTC m=+488.432546215" Mar 20 11:05:17 crc kubenswrapper[4846]: I0320 11:05:17.087326 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zb6fs" podStartSLOduration=2.647819365 podStartE2EDuration="5.087300478s" podCreationTimestamp="2026-03-20 11:05:12 +0000 UTC" firstStartedPulling="2026-03-20 11:05:13.971152047 +0000 UTC m=+485.337560293" lastFinishedPulling="2026-03-20 11:05:16.41063317 +0000 UTC m=+487.777041406" observedRunningTime="2026-03-20 11:05:17.082341456 +0000 UTC m=+488.448749702" watchObservedRunningTime="2026-03-20 11:05:17.087300478 +0000 UTC m=+488.453708714" Mar 20 11:05:18 crc kubenswrapper[4846]: I0320 11:05:18.048755 4846 generic.go:334] "Generic (PLEG): container finished" podID="70703261-65d2-40da-aa0d-948663787c75" containerID="83c841a46d124af75ff21bd9cd6a2d97b38ee56aca8894ccac30191b78fcc7b4" exitCode=0 Mar 20 11:05:18 crc kubenswrapper[4846]: I0320 11:05:18.049193 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcsxf" event={"ID":"70703261-65d2-40da-aa0d-948663787c75","Type":"ContainerDied","Data":"83c841a46d124af75ff21bd9cd6a2d97b38ee56aca8894ccac30191b78fcc7b4"} Mar 20 11:05:18 crc kubenswrapper[4846]: I0320 11:05:18.053214 4846 generic.go:334] "Generic (PLEG): container finished" podID="30a01aab-f965-4b19-97b5-a3bccd0aaff3" containerID="31949e993aa46202feee5fe874c0c4cbb45745bd20187d84e5310007bbd986ce" exitCode=0 Mar 20 11:05:18 crc kubenswrapper[4846]: I0320 11:05:18.053396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqpzh" event={"ID":"30a01aab-f965-4b19-97b5-a3bccd0aaff3","Type":"ContainerDied","Data":"31949e993aa46202feee5fe874c0c4cbb45745bd20187d84e5310007bbd986ce"} Mar 20 11:05:19 crc kubenswrapper[4846]: I0320 11:05:19.069343 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcsxf" event={"ID":"70703261-65d2-40da-aa0d-948663787c75","Type":"ContainerStarted","Data":"e7f7a7befce33cb4b49d4d19032a479dda23e0495d96038bf6337f7e41cac0a8"} Mar 20 11:05:19 crc kubenswrapper[4846]: I0320 11:05:19.073880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nqpzh" event={"ID":"30a01aab-f965-4b19-97b5-a3bccd0aaff3","Type":"ContainerStarted","Data":"a98a5fa2d6f8bdc1027bb4cab9e716bf09da72b968dcfdd24369fec784ff1fb8"} Mar 20 11:05:19 crc kubenswrapper[4846]: I0320 11:05:19.100210 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bcsxf" podStartSLOduration=1.525970902 podStartE2EDuration="4.100180146s" podCreationTimestamp="2026-03-20 11:05:15 +0000 UTC" firstStartedPulling="2026-03-20 11:05:15.996525149 +0000 UTC m=+487.362933395" lastFinishedPulling="2026-03-20 11:05:18.570734393 +0000 UTC m=+489.937142639" observedRunningTime="2026-03-20 11:05:19.096239867 +0000 UTC m=+490.462648123" watchObservedRunningTime="2026-03-20 11:05:19.100180146 +0000 UTC m=+490.466588392" Mar 20 11:05:19 crc kubenswrapper[4846]: I0320 11:05:19.128880 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nqpzh" podStartSLOduration=2.684422565 podStartE2EDuration="4.128850925s" podCreationTimestamp="2026-03-20 11:05:15 +0000 UTC" firstStartedPulling="2026-03-20 11:05:17.029595121 +0000 UTC m=+488.396003357" lastFinishedPulling="2026-03-20 11:05:18.474023471 +0000 UTC m=+489.840431717" observedRunningTime="2026-03-20 11:05:19.121964289 +0000 UTC m=+490.488372575" watchObservedRunningTime="2026-03-20 11:05:19.128850925 +0000 UTC m=+490.495259171" Mar 20 11:05:22 crc kubenswrapper[4846]: I0320 11:05:22.988537 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:22 crc kubenswrapper[4846]: I0320 11:05:22.988943 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:23 crc kubenswrapper[4846]: I0320 11:05:23.046509 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:23 crc kubenswrapper[4846]: I0320 11:05:23.146130 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mkqk" Mar 20 11:05:23 crc kubenswrapper[4846]: I0320 11:05:23.192661 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:23 crc kubenswrapper[4846]: I0320 11:05:23.192746 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:23 crc kubenswrapper[4846]: I0320 11:05:23.239513 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:24 crc kubenswrapper[4846]: I0320 11:05:24.155535 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:05:25 crc kubenswrapper[4846]: I0320 11:05:25.450930 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:25 crc kubenswrapper[4846]: I0320 11:05:25.451005 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:25 crc kubenswrapper[4846]: I0320 11:05:25.502020 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:25 crc kubenswrapper[4846]: I0320 11:05:25.621069 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:25 crc kubenswrapper[4846]: I0320 11:05:25.621133 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:26 crc kubenswrapper[4846]: I0320 11:05:26.156197 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nqpzh" Mar 20 11:05:26 crc kubenswrapper[4846]: I0320 11:05:26.437863 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" podUID="a9392c91-350e-4052-a357-5c4da7651485" containerName="registry" containerID="cri-o://c0d3b3a8cdab4ab5bd7410f5ee6afdd8a0df9d340cf87802b2110240959c0602" gracePeriod=30 Mar 20 11:05:26 crc kubenswrapper[4846]: I0320 11:05:26.660290 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bcsxf" podUID="70703261-65d2-40da-aa0d-948663787c75" containerName="registry-server" probeResult="failure" output=< Mar 20 11:05:26 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:05:26 crc kubenswrapper[4846]: > Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.122723 4846 generic.go:334] "Generic (PLEG): container finished" podID="a9392c91-350e-4052-a357-5c4da7651485" containerID="c0d3b3a8cdab4ab5bd7410f5ee6afdd8a0df9d340cf87802b2110240959c0602" exitCode=0 Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.122794 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" event={"ID":"a9392c91-350e-4052-a357-5c4da7651485","Type":"ContainerDied","Data":"c0d3b3a8cdab4ab5bd7410f5ee6afdd8a0df9d340cf87802b2110240959c0602"} Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.567456 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.635205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.635506 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.635560 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.635590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ndml\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.635625 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.636551 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.636613 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.636694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted\") pod \"a9392c91-350e-4052-a357-5c4da7651485\" (UID: \"a9392c91-350e-4052-a357-5c4da7651485\") " Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.637064 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.638219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.646124 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.652860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.652984 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml" (OuterVolumeSpecName: "kube-api-access-7ndml") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "kube-api-access-7ndml". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.654293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.657077 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.663076 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a9392c91-350e-4052-a357-5c4da7651485" (UID: "a9392c91-350e-4052-a357-5c4da7651485"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738051 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738092 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ndml\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-kube-api-access-7ndml\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738105 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9392c91-350e-4052-a357-5c4da7651485-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738122 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738135 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9392c91-350e-4052-a357-5c4da7651485-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738147 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9392c91-350e-4052-a357-5c4da7651485-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:27 crc kubenswrapper[4846]: I0320 11:05:27.738159 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9392c91-350e-4052-a357-5c4da7651485-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 11:05:28 crc kubenswrapper[4846]: I0320 11:05:28.133308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" event={"ID":"a9392c91-350e-4052-a357-5c4da7651485","Type":"ContainerDied","Data":"380f0f7bc5484ec624cc8d631e7da172a1ac3fb95607581cceaec1f6ec41a089"} Mar 20 11:05:28 crc kubenswrapper[4846]: I0320 11:05:28.133412 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cn4z4" Mar 20 11:05:28 crc kubenswrapper[4846]: I0320 11:05:28.133688 4846 scope.go:117] "RemoveContainer" containerID="c0d3b3a8cdab4ab5bd7410f5ee6afdd8a0df9d340cf87802b2110240959c0602" Mar 20 11:05:28 crc kubenswrapper[4846]: I0320 11:05:28.186948 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:05:28 crc kubenswrapper[4846]: I0320 11:05:28.193316 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cn4z4"] Mar 20 11:05:29 crc kubenswrapper[4846]: I0320 11:05:29.331369 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9392c91-350e-4052-a357-5c4da7651485" path="/var/lib/kubelet/pods/a9392c91-350e-4052-a357-5c4da7651485/volumes" Mar 20 11:05:35 crc kubenswrapper[4846]: I0320 11:05:35.664799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:35 crc kubenswrapper[4846]: I0320 11:05:35.716829 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bcsxf" Mar 20 11:05:39 crc kubenswrapper[4846]: I0320 11:05:39.677735 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:05:39 crc kubenswrapper[4846]: I0320 11:05:39.678321 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:05:39 crc kubenswrapper[4846]: I0320 11:05:39.678392 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:05:39 crc kubenswrapper[4846]: I0320 11:05:39.679192 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:05:39 crc kubenswrapper[4846]: I0320 11:05:39.679255 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c" gracePeriod=600 Mar 20 11:05:40 crc kubenswrapper[4846]: I0320 11:05:40.223800 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c" exitCode=0 Mar 20 11:05:40 crc kubenswrapper[4846]: I0320 11:05:40.223943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c"} Mar 20 11:05:40 crc kubenswrapper[4846]: I0320 11:05:40.224757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7"} Mar 20 11:05:40 crc kubenswrapper[4846]: I0320 11:05:40.224801 4846 scope.go:117] "RemoveContainer" containerID="1923a499ca89fbe04978a4722f4490935981c6541287900afe960d90049bc22e" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.148948 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566746-fkvh8"] Mar 20 11:06:00 crc kubenswrapper[4846]: E0320 11:06:00.149844 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9392c91-350e-4052-a357-5c4da7651485" containerName="registry" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.149858 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9392c91-350e-4052-a357-5c4da7651485" containerName="registry" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.149976 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9392c91-350e-4052-a357-5c4da7651485" containerName="registry" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.153443 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.155720 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.156493 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.156604 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.159936 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566746-fkvh8"] Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.217137 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvwg\" (UniqueName: \"kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg\") pod \"auto-csr-approver-29566746-fkvh8\" (UID: \"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3\") " pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.319251 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvwg\" (UniqueName: \"kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg\") pod \"auto-csr-approver-29566746-fkvh8\" (UID: \"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3\") " pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.341921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvwg\" (UniqueName: \"kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg\") pod \"auto-csr-approver-29566746-fkvh8\" (UID: \"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3\") " pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.479247 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.686378 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566746-fkvh8"] Mar 20 11:06:00 crc kubenswrapper[4846]: I0320 11:06:00.694566 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:06:01 crc kubenswrapper[4846]: I0320 11:06:01.374516 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" event={"ID":"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3","Type":"ContainerStarted","Data":"2e9648fda977e0b7ce7b8ee87ea0a13498d5a069600e7b2c1bda92a311c8ff36"} Mar 20 11:06:03 crc kubenswrapper[4846]: I0320 11:06:03.400246 4846 generic.go:334] "Generic (PLEG): container finished" podID="5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" containerID="a8a327098bfd4307ec579af79ebc7d1cc71cf40b1512bdcfc9616765affb0a9f" exitCode=0 Mar 20 11:06:03 crc kubenswrapper[4846]: I0320 11:06:03.400503 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" event={"ID":"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3","Type":"ContainerDied","Data":"a8a327098bfd4307ec579af79ebc7d1cc71cf40b1512bdcfc9616765affb0a9f"} Mar 20 11:06:04 crc kubenswrapper[4846]: I0320 11:06:04.681479 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:04 crc kubenswrapper[4846]: I0320 11:06:04.792776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hvwg\" (UniqueName: \"kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg\") pod \"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3\" (UID: \"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3\") " Mar 20 11:06:04 crc kubenswrapper[4846]: I0320 11:06:04.802123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg" (OuterVolumeSpecName: "kube-api-access-2hvwg") pod "5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" (UID: "5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3"). InnerVolumeSpecName "kube-api-access-2hvwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:06:04 crc kubenswrapper[4846]: I0320 11:06:04.894743 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hvwg\" (UniqueName: \"kubernetes.io/projected/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3-kube-api-access-2hvwg\") on node \"crc\" DevicePath \"\"" Mar 20 11:06:05 crc kubenswrapper[4846]: I0320 11:06:05.424138 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" event={"ID":"5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3","Type":"ContainerDied","Data":"2e9648fda977e0b7ce7b8ee87ea0a13498d5a069600e7b2c1bda92a311c8ff36"} Mar 20 11:06:05 crc kubenswrapper[4846]: I0320 11:06:05.424665 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e9648fda977e0b7ce7b8ee87ea0a13498d5a069600e7b2c1bda92a311c8ff36" Mar 20 11:06:05 crc kubenswrapper[4846]: I0320 11:06:05.424232 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566746-fkvh8" Mar 20 11:06:05 crc kubenswrapper[4846]: I0320 11:06:05.746600 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566740-46wjt"] Mar 20 11:06:05 crc kubenswrapper[4846]: I0320 11:06:05.749549 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566740-46wjt"] Mar 20 11:06:07 crc kubenswrapper[4846]: I0320 11:06:07.332507 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dd31859-c683-4f1a-b167-3f20a96b6f9c" path="/var/lib/kubelet/pods/8dd31859-c683-4f1a-b167-3f20a96b6f9c/volumes" Mar 20 11:07:13 crc kubenswrapper[4846]: I0320 11:07:13.233878 4846 scope.go:117] "RemoveContainer" containerID="217dee6b0998b8db43f8127c84e961c5451fee355873d531123fc419c35d02a1" Mar 20 11:07:13 crc kubenswrapper[4846]: I0320 11:07:13.259220 4846 scope.go:117] "RemoveContainer" containerID="018ae237d92648099fca8b5006a1637688b0e36b99650e00d7aab8ead8d83434" Mar 20 11:07:39 crc kubenswrapper[4846]: I0320 11:07:39.678424 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:07:39 crc kubenswrapper[4846]: I0320 11:07:39.679308 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.150718 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566748-9t2fn"] Mar 20 11:08:00 crc kubenswrapper[4846]: E0320 11:08:00.152016 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" containerName="oc" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.152040 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" containerName="oc" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.152203 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" containerName="oc" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.152964 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.156319 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.157010 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.157533 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.158492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566748-9t2fn"] Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.238076 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kg95\" (UniqueName: \"kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95\") pod \"auto-csr-approver-29566748-9t2fn\" (UID: \"456687eb-df67-4f43-81ef-4c7d146cc813\") " pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.340009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kg95\" (UniqueName: \"kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95\") pod \"auto-csr-approver-29566748-9t2fn\" (UID: \"456687eb-df67-4f43-81ef-4c7d146cc813\") " pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.366753 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kg95\" (UniqueName: \"kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95\") pod \"auto-csr-approver-29566748-9t2fn\" (UID: \"456687eb-df67-4f43-81ef-4c7d146cc813\") " pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.476400 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.926144 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566748-9t2fn"] Mar 20 11:08:00 crc kubenswrapper[4846]: I0320 11:08:00.978456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" event={"ID":"456687eb-df67-4f43-81ef-4c7d146cc813","Type":"ContainerStarted","Data":"669aaafb34e1e64bf536b8ddc1459a92a1066b1e7864255fbd119e5296eafb29"} Mar 20 11:08:02 crc kubenswrapper[4846]: I0320 11:08:02.994428 4846 generic.go:334] "Generic (PLEG): container finished" podID="456687eb-df67-4f43-81ef-4c7d146cc813" containerID="243e667f4ac6502ea55aaa409785e02a174f6d47feaa7a7767a7546bcdd4f6d8" exitCode=0 Mar 20 11:08:02 crc kubenswrapper[4846]: I0320 11:08:02.994994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" event={"ID":"456687eb-df67-4f43-81ef-4c7d146cc813","Type":"ContainerDied","Data":"243e667f4ac6502ea55aaa409785e02a174f6d47feaa7a7767a7546bcdd4f6d8"} Mar 20 11:08:04 crc kubenswrapper[4846]: I0320 11:08:04.307539 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:04 crc kubenswrapper[4846]: I0320 11:08:04.499485 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kg95\" (UniqueName: \"kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95\") pod \"456687eb-df67-4f43-81ef-4c7d146cc813\" (UID: \"456687eb-df67-4f43-81ef-4c7d146cc813\") " Mar 20 11:08:04 crc kubenswrapper[4846]: I0320 11:08:04.510331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95" (OuterVolumeSpecName: "kube-api-access-5kg95") pod "456687eb-df67-4f43-81ef-4c7d146cc813" (UID: "456687eb-df67-4f43-81ef-4c7d146cc813"). InnerVolumeSpecName "kube-api-access-5kg95". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:08:04 crc kubenswrapper[4846]: I0320 11:08:04.601981 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kg95\" (UniqueName: \"kubernetes.io/projected/456687eb-df67-4f43-81ef-4c7d146cc813-kube-api-access-5kg95\") on node \"crc\" DevicePath \"\"" Mar 20 11:08:05 crc kubenswrapper[4846]: I0320 11:08:05.013289 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" event={"ID":"456687eb-df67-4f43-81ef-4c7d146cc813","Type":"ContainerDied","Data":"669aaafb34e1e64bf536b8ddc1459a92a1066b1e7864255fbd119e5296eafb29"} Mar 20 11:08:05 crc kubenswrapper[4846]: I0320 11:08:05.013342 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566748-9t2fn" Mar 20 11:08:05 crc kubenswrapper[4846]: I0320 11:08:05.013379 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="669aaafb34e1e64bf536b8ddc1459a92a1066b1e7864255fbd119e5296eafb29" Mar 20 11:08:05 crc kubenswrapper[4846]: I0320 11:08:05.381218 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566742-wqk8p"] Mar 20 11:08:05 crc kubenswrapper[4846]: I0320 11:08:05.385573 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566742-wqk8p"] Mar 20 11:08:07 crc kubenswrapper[4846]: I0320 11:08:07.329960 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe1698aa-b5c6-4480-a886-c89481d08c59" path="/var/lib/kubelet/pods/fe1698aa-b5c6-4480-a886-c89481d08c59/volumes" Mar 20 11:08:09 crc kubenswrapper[4846]: I0320 11:08:09.678522 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:08:09 crc kubenswrapper[4846]: I0320 11:08:09.678619 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.297056 4846 scope.go:117] "RemoveContainer" containerID="4bfa365c6112a2b52c8b380c45beec1408fd692036f0d3fd40c624ec3b03310e" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.347064 4846 scope.go:117] "RemoveContainer" containerID="a3eacf9b9ba175458f62f229ad9be0b1beb42084de63bb452faf0dbc5168f703" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.397018 4846 scope.go:117] "RemoveContainer" containerID="7694dc1764c7328d2c31062299e1d6cd0fe863b3ba7083df719e4d4d21900533" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.420694 4846 scope.go:117] "RemoveContainer" containerID="39f3bd4df3b33bcdda9be40c49a4db3ea19048c00f9fff457eac20baa3ff7677" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.446423 4846 scope.go:117] "RemoveContainer" containerID="654e75bc0015d51a0a4f8cd4fc259f6c082039319f43ee1f2e05d520b54a8d47" Mar 20 11:08:13 crc kubenswrapper[4846]: I0320 11:08:13.467756 4846 scope.go:117] "RemoveContainer" containerID="3d08ab3d8c6a5543e359243a4c72f68ce419280ab30fd73f67f7e8bf9403c88f" Mar 20 11:08:39 crc kubenswrapper[4846]: I0320 11:08:39.678282 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:08:39 crc kubenswrapper[4846]: I0320 11:08:39.680182 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:08:39 crc kubenswrapper[4846]: I0320 11:08:39.680296 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:08:39 crc kubenswrapper[4846]: I0320 11:08:39.681346 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:08:39 crc kubenswrapper[4846]: I0320 11:08:39.681444 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7" gracePeriod=600 Mar 20 11:08:40 crc kubenswrapper[4846]: I0320 11:08:40.264111 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7" exitCode=0 Mar 20 11:08:40 crc kubenswrapper[4846]: I0320 11:08:40.264642 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7"} Mar 20 11:08:40 crc kubenswrapper[4846]: I0320 11:08:40.264675 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57"} Mar 20 11:08:40 crc kubenswrapper[4846]: I0320 11:08:40.264695 4846 scope.go:117] "RemoveContainer" containerID="54eaaee427f5f483944ce74b4027b7e56fabe1f2d807cd83e74c2e75873bc29c" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.143575 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566750-dfpjf"] Mar 20 11:10:00 crc kubenswrapper[4846]: E0320 11:10:00.144789 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="456687eb-df67-4f43-81ef-4c7d146cc813" containerName="oc" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.144805 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="456687eb-df67-4f43-81ef-4c7d146cc813" containerName="oc" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.144924 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="456687eb-df67-4f43-81ef-4c7d146cc813" containerName="oc" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.145435 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.148361 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.148594 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.148645 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.148823 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566750-dfpjf"] Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.283356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp6vn\" (UniqueName: \"kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn\") pod \"auto-csr-approver-29566750-dfpjf\" (UID: \"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f\") " pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.384996 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp6vn\" (UniqueName: \"kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn\") pod \"auto-csr-approver-29566750-dfpjf\" (UID: \"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f\") " pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.405569 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp6vn\" (UniqueName: \"kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn\") pod \"auto-csr-approver-29566750-dfpjf\" (UID: \"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f\") " pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.465885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.659755 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566750-dfpjf"] Mar 20 11:10:00 crc kubenswrapper[4846]: I0320 11:10:00.845160 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" event={"ID":"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f","Type":"ContainerStarted","Data":"1fc434f72cd2788ad75076451185b305b723d3642ed94f7e25ff439a0e722b6e"} Mar 20 11:10:01 crc kubenswrapper[4846]: I0320 11:10:01.856220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" event={"ID":"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f","Type":"ContainerStarted","Data":"72b4af7ccff029fbd1ac3121efe9a09c2853adaf8f43939366dda430bd481890"} Mar 20 11:10:02 crc kubenswrapper[4846]: I0320 11:10:02.863409 4846 generic.go:334] "Generic (PLEG): container finished" podID="a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" containerID="72b4af7ccff029fbd1ac3121efe9a09c2853adaf8f43939366dda430bd481890" exitCode=0 Mar 20 11:10:02 crc kubenswrapper[4846]: I0320 11:10:02.863473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" event={"ID":"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f","Type":"ContainerDied","Data":"72b4af7ccff029fbd1ac3121efe9a09c2853adaf8f43939366dda430bd481890"} Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.093589 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.237010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp6vn\" (UniqueName: \"kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn\") pod \"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f\" (UID: \"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f\") " Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.243909 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn" (OuterVolumeSpecName: "kube-api-access-pp6vn") pod "a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" (UID: "a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f"). InnerVolumeSpecName "kube-api-access-pp6vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.338700 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp6vn\" (UniqueName: \"kubernetes.io/projected/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f-kube-api-access-pp6vn\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.882703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" event={"ID":"a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f","Type":"ContainerDied","Data":"1fc434f72cd2788ad75076451185b305b723d3642ed94f7e25ff439a0e722b6e"} Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.882757 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fc434f72cd2788ad75076451185b305b723d3642ed94f7e25ff439a0e722b6e" Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.882799 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566750-dfpjf" Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.936397 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566744-n8dcz"] Mar 20 11:10:04 crc kubenswrapper[4846]: I0320 11:10:04.944090 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566744-n8dcz"] Mar 20 11:10:05 crc kubenswrapper[4846]: I0320 11:10:05.332039 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf95362a-3893-4688-8c22-7092d445382a" path="/var/lib/kubelet/pods/cf95362a-3893-4688-8c22-7092d445382a/volumes" Mar 20 11:10:13 crc kubenswrapper[4846]: I0320 11:10:13.574869 4846 scope.go:117] "RemoveContainer" containerID="7722d1bc9a0f5a66664d96d60cd3cc0f8782115391203bbf8bffd025c897c5a6" Mar 20 11:10:13 crc kubenswrapper[4846]: I0320 11:10:13.598139 4846 scope.go:117] "RemoveContainer" containerID="347a272e87ec1fb8bd4cea9f378d7adf936795fbc3ab3bae6eaf68fa9afa595a" Mar 20 11:10:39 crc kubenswrapper[4846]: I0320 11:10:39.678042 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:10:39 crc kubenswrapper[4846]: I0320 11:10:39.678702 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.741306 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-dwsx8"] Mar 20 11:10:43 crc kubenswrapper[4846]: E0320 11:10:43.742132 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" containerName="oc" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.742153 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" containerName="oc" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.742314 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" containerName="oc" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.742970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-dwsx8" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.746035 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq"] Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.747295 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.747758 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.748856 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-j2snt" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.749072 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.753137 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq"] Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.755444 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jj95g" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.758708 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-dwsx8"] Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.789226 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-dcs94"] Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.790493 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.793838 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-868wt" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.812837 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-dcs94"] Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.833349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88s5w\" (UniqueName: \"kubernetes.io/projected/9480649a-6163-4ab6-90be-d825b1d39724-kube-api-access-88s5w\") pod \"cert-manager-cainjector-cf98fcc89-xwvmq\" (UID: \"9480649a-6163-4ab6-90be-d825b1d39724\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.833437 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mptb\" (UniqueName: \"kubernetes.io/projected/64fa0d45-f6a9-4b1f-8381-8d0cf63780e9-kube-api-access-7mptb\") pod \"cert-manager-858654f9db-dwsx8\" (UID: \"64fa0d45-f6a9-4b1f-8381-8d0cf63780e9\") " pod="cert-manager/cert-manager-858654f9db-dwsx8" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.833496 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6lwk\" (UniqueName: \"kubernetes.io/projected/0794f7a2-b0b8-419a-b7db-eb75a2dcbf73-kube-api-access-f6lwk\") pod \"cert-manager-webhook-687f57d79b-dcs94\" (UID: \"0794f7a2-b0b8-419a-b7db-eb75a2dcbf73\") " pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.934733 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mptb\" (UniqueName: \"kubernetes.io/projected/64fa0d45-f6a9-4b1f-8381-8d0cf63780e9-kube-api-access-7mptb\") pod \"cert-manager-858654f9db-dwsx8\" (UID: \"64fa0d45-f6a9-4b1f-8381-8d0cf63780e9\") " pod="cert-manager/cert-manager-858654f9db-dwsx8" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.935330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6lwk\" (UniqueName: \"kubernetes.io/projected/0794f7a2-b0b8-419a-b7db-eb75a2dcbf73-kube-api-access-f6lwk\") pod \"cert-manager-webhook-687f57d79b-dcs94\" (UID: \"0794f7a2-b0b8-419a-b7db-eb75a2dcbf73\") " pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.935368 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88s5w\" (UniqueName: \"kubernetes.io/projected/9480649a-6163-4ab6-90be-d825b1d39724-kube-api-access-88s5w\") pod \"cert-manager-cainjector-cf98fcc89-xwvmq\" (UID: \"9480649a-6163-4ab6-90be-d825b1d39724\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.967127 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88s5w\" (UniqueName: \"kubernetes.io/projected/9480649a-6163-4ab6-90be-d825b1d39724-kube-api-access-88s5w\") pod \"cert-manager-cainjector-cf98fcc89-xwvmq\" (UID: \"9480649a-6163-4ab6-90be-d825b1d39724\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.967690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mptb\" (UniqueName: \"kubernetes.io/projected/64fa0d45-f6a9-4b1f-8381-8d0cf63780e9-kube-api-access-7mptb\") pod \"cert-manager-858654f9db-dwsx8\" (UID: \"64fa0d45-f6a9-4b1f-8381-8d0cf63780e9\") " pod="cert-manager/cert-manager-858654f9db-dwsx8" Mar 20 11:10:43 crc kubenswrapper[4846]: I0320 11:10:43.978422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6lwk\" (UniqueName: \"kubernetes.io/projected/0794f7a2-b0b8-419a-b7db-eb75a2dcbf73-kube-api-access-f6lwk\") pod \"cert-manager-webhook-687f57d79b-dcs94\" (UID: \"0794f7a2-b0b8-419a-b7db-eb75a2dcbf73\") " pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.073387 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-dwsx8" Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.081585 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.110060 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.313448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-dwsx8"] Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.340747 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq"] Mar 20 11:10:44 crc kubenswrapper[4846]: W0320 11:10:44.350462 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9480649a_6163_4ab6_90be_d825b1d39724.slice/crio-2f39f693c1e750075e3bdd37878013ecdbe9472063a58726818fd1b61482b8e8 WatchSource:0}: Error finding container 2f39f693c1e750075e3bdd37878013ecdbe9472063a58726818fd1b61482b8e8: Status 404 returned error can't find the container with id 2f39f693c1e750075e3bdd37878013ecdbe9472063a58726818fd1b61482b8e8 Mar 20 11:10:44 crc kubenswrapper[4846]: I0320 11:10:44.376213 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-dcs94"] Mar 20 11:10:44 crc kubenswrapper[4846]: W0320 11:10:44.378664 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0794f7a2_b0b8_419a_b7db_eb75a2dcbf73.slice/crio-9291dc88a2b46c30f1089b481b37ffac0751e8d4386298bb213672679a06e895 WatchSource:0}: Error finding container 9291dc88a2b46c30f1089b481b37ffac0751e8d4386298bb213672679a06e895: Status 404 returned error can't find the container with id 9291dc88a2b46c30f1089b481b37ffac0751e8d4386298bb213672679a06e895 Mar 20 11:10:45 crc kubenswrapper[4846]: I0320 11:10:45.139811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" event={"ID":"0794f7a2-b0b8-419a-b7db-eb75a2dcbf73","Type":"ContainerStarted","Data":"9291dc88a2b46c30f1089b481b37ffac0751e8d4386298bb213672679a06e895"} Mar 20 11:10:45 crc kubenswrapper[4846]: I0320 11:10:45.140705 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" event={"ID":"9480649a-6163-4ab6-90be-d825b1d39724","Type":"ContainerStarted","Data":"2f39f693c1e750075e3bdd37878013ecdbe9472063a58726818fd1b61482b8e8"} Mar 20 11:10:45 crc kubenswrapper[4846]: I0320 11:10:45.141479 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-dwsx8" event={"ID":"64fa0d45-f6a9-4b1f-8381-8d0cf63780e9","Type":"ContainerStarted","Data":"abbd4ad58d0a0001e7d05e1dc8da1b2ecd5e061f82e2cfaa6a65c23ef23bcc31"} Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.169140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" event={"ID":"9480649a-6163-4ab6-90be-d825b1d39724","Type":"ContainerStarted","Data":"b5ae0176e541b2de7b99b831c2ac3c055633622f129d72b38dbf863d26157c73"} Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.173721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-dwsx8" event={"ID":"64fa0d45-f6a9-4b1f-8381-8d0cf63780e9","Type":"ContainerStarted","Data":"d922c627aded33c29d0ec44b43a99e0ef91e66f2ea0b0f5c4492f7aa8e29e2d6"} Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.175581 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" event={"ID":"0794f7a2-b0b8-419a-b7db-eb75a2dcbf73","Type":"ContainerStarted","Data":"dfc666222de77e373eb376598858f74274461fd778100d60dc587066026a6201"} Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.176222 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.200613 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xwvmq" podStartSLOduration=2.377144555 podStartE2EDuration="6.20058238s" podCreationTimestamp="2026-03-20 11:10:43 +0000 UTC" firstStartedPulling="2026-03-20 11:10:44.354805343 +0000 UTC m=+815.721213579" lastFinishedPulling="2026-03-20 11:10:48.178243168 +0000 UTC m=+819.544651404" observedRunningTime="2026-03-20 11:10:49.189284763 +0000 UTC m=+820.555692999" watchObservedRunningTime="2026-03-20 11:10:49.20058238 +0000 UTC m=+820.566990616" Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.223707 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" podStartSLOduration=2.358286241 podStartE2EDuration="6.223684956s" podCreationTimestamp="2026-03-20 11:10:43 +0000 UTC" firstStartedPulling="2026-03-20 11:10:44.381530768 +0000 UTC m=+815.747939004" lastFinishedPulling="2026-03-20 11:10:48.246929483 +0000 UTC m=+819.613337719" observedRunningTime="2026-03-20 11:10:49.215364682 +0000 UTC m=+820.581773058" watchObservedRunningTime="2026-03-20 11:10:49.223684956 +0000 UTC m=+820.590093192" Mar 20 11:10:49 crc kubenswrapper[4846]: I0320 11:10:49.235967 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-dwsx8" podStartSLOduration=2.396486567 podStartE2EDuration="6.235946577s" podCreationTimestamp="2026-03-20 11:10:43 +0000 UTC" firstStartedPulling="2026-03-20 11:10:44.338750418 +0000 UTC m=+815.705158654" lastFinishedPulling="2026-03-20 11:10:48.178210428 +0000 UTC m=+819.544618664" observedRunningTime="2026-03-20 11:10:49.235350213 +0000 UTC m=+820.601758459" watchObservedRunningTime="2026-03-20 11:10:49.235946577 +0000 UTC m=+820.602354813" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.099679 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqml5"] Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100538 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-controller" containerID="cri-o://88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100602 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-node" containerID="cri-o://a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100636 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="northd" containerID="cri-o://39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100678 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="sbdb" containerID="cri-o://af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100708 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="nbdb" containerID="cri-o://446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100679 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-acl-logging" containerID="cri-o://996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.100705 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.172567 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" containerID="cri-o://57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" gracePeriod=30 Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.537735 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/3.log" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.539930 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovn-acl-logging/0.log" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.540454 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovn-controller/0.log" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.540885 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.592973 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mmbdx"] Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593272 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593296 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593311 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kubecfg-setup" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593318 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kubecfg-setup" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593326 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593333 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593347 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593354 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593363 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593370 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593382 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="sbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593390 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="sbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593401 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-acl-logging" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593410 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-acl-logging" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593425 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593433 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593442 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="nbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593449 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="nbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593458 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-node" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593466 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-node" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593476 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="northd" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593484 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="northd" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593617 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="nbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593635 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593647 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593657 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-node" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593668 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="sbdb" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593678 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-acl-logging" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593690 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="northd" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593700 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovn-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593711 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593721 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593860 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593870 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: E0320 11:10:53.593885 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.593913 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.594107 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.594393 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerName="ovnkube-controller" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.596182 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603640 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603747 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.603954 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604018 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604061 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv8k6\" (UniqueName: \"kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604089 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604119 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604161 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604187 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604218 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604270 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd\") pod \"d1a81831-258d-4fad-adc4-02dee4e744d7\" (UID: \"d1a81831-258d-4fad-adc4-02dee4e744d7\") " Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604590 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket" (OuterVolumeSpecName: "log-socket") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604873 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604910 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604923 4846 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-log-socket\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604951 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash" (OuterVolumeSpecName: "host-slash") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.604976 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log" (OuterVolumeSpecName: "node-log") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605296 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605322 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605345 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605413 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605462 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605457 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605510 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.605627 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.612348 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.614400 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6" (OuterVolumeSpecName: "kube-api-access-bv8k6") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "kube-api-access-bv8k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.623673 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d1a81831-258d-4fad-adc4-02dee4e744d7" (UID: "d1a81831-258d-4fad-adc4-02dee4e744d7"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706137 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqm28\" (UniqueName: \"kubernetes.io/projected/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-kube-api-access-xqm28\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovn-node-metrics-cert\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-netns\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706341 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-slash\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706367 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-var-lib-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706387 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-config\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-kubelet\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-log-socket\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706601 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-ovn\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-bin\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706746 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706786 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-node-log\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.706860 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-systemd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-script-lib\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707094 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-systemd-units\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707133 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-env-overrides\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-netd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-etc-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707291 4846 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707314 4846 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707329 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707341 4846 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707352 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707368 4846 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707382 4846 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-slash\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707396 4846 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707409 4846 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-node-log\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707421 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707432 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1a81831-258d-4fad-adc4-02dee4e744d7-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707444 4846 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707455 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1a81831-258d-4fad-adc4-02dee4e744d7-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707465 4846 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707476 4846 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707490 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv8k6\" (UniqueName: \"kubernetes.io/projected/d1a81831-258d-4fad-adc4-02dee4e744d7-kube-api-access-bv8k6\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.707503 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1a81831-258d-4fad-adc4-02dee4e744d7-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809220 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqm28\" (UniqueName: \"kubernetes.io/projected/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-kube-api-access-xqm28\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovn-node-metrics-cert\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-netns\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-slash\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-var-lib-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-config\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809561 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809613 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-slash\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809583 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-kubelet\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-var-lib-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-log-socket\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809641 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-run-netns\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-log-socket\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-kubelet\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-ovn\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-ovn\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809817 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-bin\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809859 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-bin\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-node-log\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-node-log\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.809935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-systemd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-systemd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810102 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-run-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-script-lib\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-systemd-units\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810248 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-env-overrides\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-etc-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810303 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-systemd-units\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810330 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-netd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-host-cni-netd\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-etc-openvswitch\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-config\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.810830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovnkube-script-lib\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.811033 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-env-overrides\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.814658 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-ovn-node-metrics-cert\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.831280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqm28\" (UniqueName: \"kubernetes.io/projected/4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1-kube-api-access-xqm28\") pod \"ovnkube-node-mmbdx\" (UID: \"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1\") " pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:53 crc kubenswrapper[4846]: I0320 11:10:53.922563 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.113390 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-dcs94" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.219553 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/2.log" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.220258 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/1.log" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.220349 4846 generic.go:334] "Generic (PLEG): container finished" podID="dfe48192-f5b6-4ba5-a4a4-475459fd1815" containerID="37bdd76cbbac0fe900855b30e99d6a6d41e56b26445376aa1a01381c1438cd84" exitCode=2 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.220428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerDied","Data":"37bdd76cbbac0fe900855b30e99d6a6d41e56b26445376aa1a01381c1438cd84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.220729 4846 scope.go:117] "RemoveContainer" containerID="8fb23cd7b49c6cc981789bce456b43e1b6b58357f86968141731fbd8f539e4f9" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.221216 4846 scope.go:117] "RemoveContainer" containerID="37bdd76cbbac0fe900855b30e99d6a6d41e56b26445376aa1a01381c1438cd84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.224658 4846 generic.go:334] "Generic (PLEG): container finished" podID="4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1" containerID="e99920bb8c36a92d57b23e2bae4031823c25cf33b1c224fac60bf9a1a0519bd8" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.224789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerDied","Data":"e99920bb8c36a92d57b23e2bae4031823c25cf33b1c224fac60bf9a1a0519bd8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.224843 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"b0e74a78613e8b19ce5d95b38da319d7f9fd1f4de80e416d1ca35c77e7dd9ed6"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.230445 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovnkube-controller/3.log" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.238750 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovn-acl-logging/0.log" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.248193 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cqml5_d1a81831-258d-4fad-adc4-02dee4e744d7/ovn-controller/0.log" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250687 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250722 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250732 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250744 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250754 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250767 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" exitCode=0 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250775 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" exitCode=143 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250784 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1a81831-258d-4fad-adc4-02dee4e744d7" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" exitCode=143 Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250890 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250944 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250960 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250967 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250974 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250982 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250988 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.250996 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251003 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251009 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251015 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251034 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251054 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251060 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251067 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251074 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251080 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251087 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251092 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251098 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251105 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251127 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251134 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251141 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251147 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251153 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251160 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251166 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251175 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251182 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251188 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251197 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" event={"ID":"d1a81831-258d-4fad-adc4-02dee4e744d7","Type":"ContainerDied","Data":"1f8b167f2df64ee51562c3a542bd0eb419c5ab206c41a98643f31e50925f9ed3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251207 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251215 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251223 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251230 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251237 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251246 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251252 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251260 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251267 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251274 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.251382 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqml5" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.314742 4846 scope.go:117] "RemoveContainer" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.343396 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.350060 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqml5"] Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.354357 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqml5"] Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.368275 4846 scope.go:117] "RemoveContainer" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.404484 4846 scope.go:117] "RemoveContainer" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.422696 4846 scope.go:117] "RemoveContainer" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.439696 4846 scope.go:117] "RemoveContainer" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.458253 4846 scope.go:117] "RemoveContainer" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.475259 4846 scope.go:117] "RemoveContainer" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.511812 4846 scope.go:117] "RemoveContainer" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.536760 4846 scope.go:117] "RemoveContainer" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.561105 4846 scope.go:117] "RemoveContainer" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.561604 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": container with ID starting with 57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114 not found: ID does not exist" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.561650 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} err="failed to get container status \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": rpc error: code = NotFound desc = could not find container \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": container with ID starting with 57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.561673 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.562340 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": container with ID starting with 28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988 not found: ID does not exist" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.562408 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} err="failed to get container status \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": rpc error: code = NotFound desc = could not find container \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": container with ID starting with 28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.562441 4846 scope.go:117] "RemoveContainer" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.562993 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": container with ID starting with af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7 not found: ID does not exist" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563020 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} err="failed to get container status \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": rpc error: code = NotFound desc = could not find container \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": container with ID starting with af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563048 4846 scope.go:117] "RemoveContainer" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.563320 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": container with ID starting with 446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84 not found: ID does not exist" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563346 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} err="failed to get container status \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": rpc error: code = NotFound desc = could not find container \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": container with ID starting with 446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563364 4846 scope.go:117] "RemoveContainer" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.563710 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": container with ID starting with 39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3 not found: ID does not exist" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563735 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} err="failed to get container status \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": rpc error: code = NotFound desc = could not find container \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": container with ID starting with 39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.563750 4846 scope.go:117] "RemoveContainer" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.564390 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": container with ID starting with cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4 not found: ID does not exist" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.564413 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} err="failed to get container status \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": rpc error: code = NotFound desc = could not find container \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": container with ID starting with cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.564427 4846 scope.go:117] "RemoveContainer" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.564648 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": container with ID starting with a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181 not found: ID does not exist" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.564668 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} err="failed to get container status \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": rpc error: code = NotFound desc = could not find container \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": container with ID starting with a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.564682 4846 scope.go:117] "RemoveContainer" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.564986 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": container with ID starting with 996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8 not found: ID does not exist" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565015 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} err="failed to get container status \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": rpc error: code = NotFound desc = could not find container \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": container with ID starting with 996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565032 4846 scope.go:117] "RemoveContainer" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.565282 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": container with ID starting with 88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a not found: ID does not exist" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565304 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} err="failed to get container status \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": rpc error: code = NotFound desc = could not find container \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": container with ID starting with 88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565316 4846 scope.go:117] "RemoveContainer" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: E0320 11:10:54.565566 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": container with ID starting with 68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec not found: ID does not exist" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565598 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} err="failed to get container status \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": rpc error: code = NotFound desc = could not find container \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": container with ID starting with 68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565619 4846 scope.go:117] "RemoveContainer" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565874 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} err="failed to get container status \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": rpc error: code = NotFound desc = could not find container \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": container with ID starting with 57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.565915 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566155 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} err="failed to get container status \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": rpc error: code = NotFound desc = could not find container \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": container with ID starting with 28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566182 4846 scope.go:117] "RemoveContainer" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566389 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} err="failed to get container status \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": rpc error: code = NotFound desc = could not find container \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": container with ID starting with af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566409 4846 scope.go:117] "RemoveContainer" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566607 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} err="failed to get container status \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": rpc error: code = NotFound desc = could not find container \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": container with ID starting with 446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566636 4846 scope.go:117] "RemoveContainer" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566848 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} err="failed to get container status \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": rpc error: code = NotFound desc = could not find container \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": container with ID starting with 39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.566869 4846 scope.go:117] "RemoveContainer" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567140 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} err="failed to get container status \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": rpc error: code = NotFound desc = could not find container \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": container with ID starting with cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567161 4846 scope.go:117] "RemoveContainer" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567445 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} err="failed to get container status \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": rpc error: code = NotFound desc = could not find container \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": container with ID starting with a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567461 4846 scope.go:117] "RemoveContainer" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567847 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} err="failed to get container status \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": rpc error: code = NotFound desc = could not find container \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": container with ID starting with 996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.567871 4846 scope.go:117] "RemoveContainer" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.568226 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} err="failed to get container status \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": rpc error: code = NotFound desc = could not find container \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": container with ID starting with 88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.568245 4846 scope.go:117] "RemoveContainer" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.568791 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} err="failed to get container status \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": rpc error: code = NotFound desc = could not find container \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": container with ID starting with 68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.568812 4846 scope.go:117] "RemoveContainer" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.569148 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} err="failed to get container status \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": rpc error: code = NotFound desc = could not find container \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": container with ID starting with 57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.569178 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.569459 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} err="failed to get container status \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": rpc error: code = NotFound desc = could not find container \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": container with ID starting with 28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.569483 4846 scope.go:117] "RemoveContainer" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.569928 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} err="failed to get container status \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": rpc error: code = NotFound desc = could not find container \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": container with ID starting with af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.570022 4846 scope.go:117] "RemoveContainer" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.570421 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} err="failed to get container status \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": rpc error: code = NotFound desc = could not find container \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": container with ID starting with 446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.570447 4846 scope.go:117] "RemoveContainer" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.570729 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} err="failed to get container status \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": rpc error: code = NotFound desc = could not find container \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": container with ID starting with 39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.570762 4846 scope.go:117] "RemoveContainer" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571142 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} err="failed to get container status \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": rpc error: code = NotFound desc = could not find container \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": container with ID starting with cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571164 4846 scope.go:117] "RemoveContainer" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571511 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} err="failed to get container status \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": rpc error: code = NotFound desc = could not find container \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": container with ID starting with a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571545 4846 scope.go:117] "RemoveContainer" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571934 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} err="failed to get container status \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": rpc error: code = NotFound desc = could not find container \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": container with ID starting with 996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.571971 4846 scope.go:117] "RemoveContainer" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.572391 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} err="failed to get container status \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": rpc error: code = NotFound desc = could not find container \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": container with ID starting with 88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.572425 4846 scope.go:117] "RemoveContainer" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.572922 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} err="failed to get container status \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": rpc error: code = NotFound desc = could not find container \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": container with ID starting with 68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.572983 4846 scope.go:117] "RemoveContainer" containerID="57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573347 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114"} err="failed to get container status \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": rpc error: code = NotFound desc = could not find container \"57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114\": container with ID starting with 57fb0a8a4cd7bd2269f720e155eeb366f681b36e2c2c48cb5506d6e37696c114 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573380 4846 scope.go:117] "RemoveContainer" containerID="28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573660 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988"} err="failed to get container status \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": rpc error: code = NotFound desc = could not find container \"28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988\": container with ID starting with 28f1c479a15febb0fe1bb5a3aae8a0a40d69f51bef685890e9716a2bb5784988 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573686 4846 scope.go:117] "RemoveContainer" containerID="af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573959 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7"} err="failed to get container status \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": rpc error: code = NotFound desc = could not find container \"af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7\": container with ID starting with af012d9e8039a5e839fcf03d6e65525b9e6f26b49e4602eaa7c4e316441d21e7 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.573984 4846 scope.go:117] "RemoveContainer" containerID="446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574211 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84"} err="failed to get container status \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": rpc error: code = NotFound desc = could not find container \"446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84\": container with ID starting with 446eecd856b8fedfe9ab9fe5d02fde1a7b845053278845d6e7cd60cecfb1ac84 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574230 4846 scope.go:117] "RemoveContainer" containerID="39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574450 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3"} err="failed to get container status \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": rpc error: code = NotFound desc = could not find container \"39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3\": container with ID starting with 39437c95be594dced85cdd97948f2991808f338bf2b39f339dba4a01c3ede1b3 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574476 4846 scope.go:117] "RemoveContainer" containerID="cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574696 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4"} err="failed to get container status \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": rpc error: code = NotFound desc = could not find container \"cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4\": container with ID starting with cbb4eb6cba105230d938a4aea44b573eb3f1cccabeba249eac6c2b2d954809a4 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574720 4846 scope.go:117] "RemoveContainer" containerID="a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574933 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181"} err="failed to get container status \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": rpc error: code = NotFound desc = could not find container \"a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181\": container with ID starting with a010d0eb0e516c0b3cf0e59acbb50519198b7bf73b5ac95c7e586541c2ea4181 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.574962 4846 scope.go:117] "RemoveContainer" containerID="996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.575190 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8"} err="failed to get container status \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": rpc error: code = NotFound desc = could not find container \"996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8\": container with ID starting with 996f5d35c328d011cd8d80ea7992c870db4c50e6e9ace5b2ccd22bec7aec39a8 not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.575219 4846 scope.go:117] "RemoveContainer" containerID="88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.575437 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a"} err="failed to get container status \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": rpc error: code = NotFound desc = could not find container \"88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a\": container with ID starting with 88a26ad43e330d3cd52cdd3b5f2253c8db64c278a9c8b542ccd9a80f2099989a not found: ID does not exist" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.575458 4846 scope.go:117] "RemoveContainer" containerID="68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec" Mar 20 11:10:54 crc kubenswrapper[4846]: I0320 11:10:54.575655 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec"} err="failed to get container status \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": rpc error: code = NotFound desc = could not find container \"68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec\": container with ID starting with 68a9d082c00060fbb39a6a3ef243f6e3c730f0d4b7ca0f4462f1b07edc1ba1ec not found: ID does not exist" Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.262036 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-cqnfg_dfe48192-f5b6-4ba5-a4a4-475459fd1815/kube-multus/2.log" Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.262528 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-cqnfg" event={"ID":"dfe48192-f5b6-4ba5-a4a4-475459fd1815","Type":"ContainerStarted","Data":"7b5cc6b0c50c7b3e2cd3002ddf0f228c3dd33f9e5244dc6ca2305d6270a7e8c1"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"5c856d99e57fbfc69168e900588fbc052f6df7f7f08650b34dfbafdc90d2215f"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"15ac7226f66aab3adc108b804cea0da818913b1af4ae5fff07256a30af5eb318"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268176 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"17bd7e488b055a63238880cb0e720353494365c7cc91ff579d714fc83e711695"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268188 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"909b76c20992455d58ee0a4d3459c995962f505bc4037c7499517dd77661fbfd"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268200 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"f7293909aa737e64aaf97bc663ddac947f7194694f339a2533ce456cb768b94a"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.268210 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"b29fb50e6702a731a75830ade95abd8948fedd43c0e77c1cc506a0eaec8f3e74"} Mar 20 11:10:55 crc kubenswrapper[4846]: I0320 11:10:55.333499 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1a81831-258d-4fad-adc4-02dee4e744d7" path="/var/lib/kubelet/pods/d1a81831-258d-4fad-adc4-02dee4e744d7/volumes" Mar 20 11:10:57 crc kubenswrapper[4846]: I0320 11:10:57.286310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"8c7bfb1bfb3d7f5374bdb09fea2fc0ca9370d8e0ce8a0d06b5dc39327ce98854"} Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.312230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" event={"ID":"4af8cbd0-d4ad-4e72-ae13-d174bd5bfbf1","Type":"ContainerStarted","Data":"057cbd7a5a6bd364898ad27994e8567547ceaa09887b8e41f98f307305908da3"} Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.312853 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.312873 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.312882 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.348626 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.349201 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" podStartSLOduration=7.349179555 podStartE2EDuration="7.349179555s" podCreationTimestamp="2026-03-20 11:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:11:00.345765181 +0000 UTC m=+831.712173407" watchObservedRunningTime="2026-03-20 11:11:00.349179555 +0000 UTC m=+831.715587791" Mar 20 11:11:00 crc kubenswrapper[4846]: I0320 11:11:00.350746 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:09 crc kubenswrapper[4846]: I0320 11:11:09.677609 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:11:09 crc kubenswrapper[4846]: I0320 11:11:09.677961 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:11:10 crc kubenswrapper[4846]: I0320 11:11:10.542703 4846 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 11:11:23 crc kubenswrapper[4846]: I0320 11:11:23.958261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mmbdx" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.394774 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7"] Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.397032 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.403380 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.406551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7"] Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.577441 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.577510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.577576 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k5bh\" (UniqueName: \"kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.679401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.679454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.679539 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k5bh\" (UniqueName: \"kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.680306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.680323 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.701207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k5bh\" (UniqueName: \"kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:34 crc kubenswrapper[4846]: I0320 11:11:34.720268 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:35 crc kubenswrapper[4846]: I0320 11:11:35.130663 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7"] Mar 20 11:11:35 crc kubenswrapper[4846]: I0320 11:11:35.524520 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerStarted","Data":"b416f1ff46115d88aca98fe1180aa955d6086200d213929e837881f05d5f71de"} Mar 20 11:11:35 crc kubenswrapper[4846]: I0320 11:11:35.526188 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerStarted","Data":"06efd58ccc30619e326ff7e71954214e9a2520e46c83c37c00d78d64f6a0cfad"} Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.451844 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.453470 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.478709 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.504619 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.504735 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.504813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp9vt\" (UniqueName: \"kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.532494 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerID="b416f1ff46115d88aca98fe1180aa955d6086200d213929e837881f05d5f71de" exitCode=0 Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.532564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerDied","Data":"b416f1ff46115d88aca98fe1180aa955d6086200d213929e837881f05d5f71de"} Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.534794 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.606748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp9vt\" (UniqueName: \"kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.606865 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.606928 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.607657 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.607649 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.633739 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp9vt\" (UniqueName: \"kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt\") pod \"redhat-operators-pxg7s\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:36 crc kubenswrapper[4846]: I0320 11:11:36.784023 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:37 crc kubenswrapper[4846]: I0320 11:11:37.012505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:37 crc kubenswrapper[4846]: I0320 11:11:37.541233 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerID="e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99" exitCode=0 Mar 20 11:11:37 crc kubenswrapper[4846]: I0320 11:11:37.541340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerDied","Data":"e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99"} Mar 20 11:11:37 crc kubenswrapper[4846]: I0320 11:11:37.541704 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerStarted","Data":"aa2264ae0db4071e9810408bef99b5514bed418d6af48d8388b36934a8b87c60"} Mar 20 11:11:38 crc kubenswrapper[4846]: I0320 11:11:38.562642 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerID="a7f9c1d93d0758ac717e2e4743ff2cb4a0d58aaeeec7e697854ac0a298d9e1a9" exitCode=0 Mar 20 11:11:38 crc kubenswrapper[4846]: I0320 11:11:38.562750 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerDied","Data":"a7f9c1d93d0758ac717e2e4743ff2cb4a0d58aaeeec7e697854ac0a298d9e1a9"} Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.574357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerStarted","Data":"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54"} Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.576654 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerID="a128dbfc489e4a7eadc0973f7933c9cb273bb15d3f008f9c3043cc9a7f2b9e5e" exitCode=0 Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.576690 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerDied","Data":"a128dbfc489e4a7eadc0973f7933c9cb273bb15d3f008f9c3043cc9a7f2b9e5e"} Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.677639 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.677746 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.677831 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.678778 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:11:39 crc kubenswrapper[4846]: I0320 11:11:39.678894 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57" gracePeriod=600 Mar 20 11:11:39 crc kubenswrapper[4846]: E0320 11:11:39.826692 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa8a1ece_2e0b_4e8a_ba55_e719e1fc333b.slice/crio-conmon-0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa8a1ece_2e0b_4e8a_ba55_e719e1fc333b.slice/crio-0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57.scope\": RecentStats: unable to find data in memory cache]" Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.586225 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerID="78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54" exitCode=0 Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.586362 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerDied","Data":"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54"} Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.590346 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57" exitCode=0 Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.590526 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57"} Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.590548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45"} Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.590567 4846 scope.go:117] "RemoveContainer" containerID="408d23d8c64f5dc49f0883204ad9724967d90b74ca1eeb6f2c6568afabfd67d7" Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.835620 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.971985 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle\") pod \"eb990abc-5080-4db8-919a-e59085f1a8b0\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.972065 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util\") pod \"eb990abc-5080-4db8-919a-e59085f1a8b0\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.972200 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k5bh\" (UniqueName: \"kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh\") pod \"eb990abc-5080-4db8-919a-e59085f1a8b0\" (UID: \"eb990abc-5080-4db8-919a-e59085f1a8b0\") " Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.973807 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle" (OuterVolumeSpecName: "bundle") pod "eb990abc-5080-4db8-919a-e59085f1a8b0" (UID: "eb990abc-5080-4db8-919a-e59085f1a8b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:11:40 crc kubenswrapper[4846]: I0320 11:11:40.980720 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh" (OuterVolumeSpecName: "kube-api-access-5k5bh") pod "eb990abc-5080-4db8-919a-e59085f1a8b0" (UID: "eb990abc-5080-4db8-919a-e59085f1a8b0"). InnerVolumeSpecName "kube-api-access-5k5bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.073737 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k5bh\" (UniqueName: \"kubernetes.io/projected/eb990abc-5080-4db8-919a-e59085f1a8b0-kube-api-access-5k5bh\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.073784 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.168651 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util" (OuterVolumeSpecName: "util") pod "eb990abc-5080-4db8-919a-e59085f1a8b0" (UID: "eb990abc-5080-4db8-919a-e59085f1a8b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.175200 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/eb990abc-5080-4db8-919a-e59085f1a8b0-util\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.600549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerStarted","Data":"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8"} Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.603967 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.603998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7" event={"ID":"eb990abc-5080-4db8-919a-e59085f1a8b0","Type":"ContainerDied","Data":"06efd58ccc30619e326ff7e71954214e9a2520e46c83c37c00d78d64f6a0cfad"} Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.604055 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06efd58ccc30619e326ff7e71954214e9a2520e46c83c37c00d78d64f6a0cfad" Mar 20 11:11:41 crc kubenswrapper[4846]: I0320 11:11:41.629037 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxg7s" podStartSLOduration=1.9800530539999999 podStartE2EDuration="5.629011257s" podCreationTimestamp="2026-03-20 11:11:36 +0000 UTC" firstStartedPulling="2026-03-20 11:11:37.543146865 +0000 UTC m=+868.909555101" lastFinishedPulling="2026-03-20 11:11:41.192105068 +0000 UTC m=+872.558513304" observedRunningTime="2026-03-20 11:11:41.623445549 +0000 UTC m=+872.989853805" watchObservedRunningTime="2026-03-20 11:11:41.629011257 +0000 UTC m=+872.995419493" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.764593 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h"] Mar 20 11:11:44 crc kubenswrapper[4846]: E0320 11:11:44.765204 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="extract" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.765218 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="extract" Mar 20 11:11:44 crc kubenswrapper[4846]: E0320 11:11:44.765230 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="util" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.765236 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="util" Mar 20 11:11:44 crc kubenswrapper[4846]: E0320 11:11:44.765256 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="pull" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.765262 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="pull" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.765371 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb990abc-5080-4db8-919a-e59085f1a8b0" containerName="extract" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.765874 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.772603 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.772608 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.772820 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-59wsm" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.776239 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h"] Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.831016 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckzls\" (UniqueName: \"kubernetes.io/projected/5db89f9f-d51b-430f-88de-34f48c693c58-kube-api-access-ckzls\") pod \"nmstate-operator-796d4cfff4-fcc6h\" (UID: \"5db89f9f-d51b-430f-88de-34f48c693c58\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.932351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckzls\" (UniqueName: \"kubernetes.io/projected/5db89f9f-d51b-430f-88de-34f48c693c58-kube-api-access-ckzls\") pod \"nmstate-operator-796d4cfff4-fcc6h\" (UID: \"5db89f9f-d51b-430f-88de-34f48c693c58\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" Mar 20 11:11:44 crc kubenswrapper[4846]: I0320 11:11:44.957962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckzls\" (UniqueName: \"kubernetes.io/projected/5db89f9f-d51b-430f-88de-34f48c693c58-kube-api-access-ckzls\") pod \"nmstate-operator-796d4cfff4-fcc6h\" (UID: \"5db89f9f-d51b-430f-88de-34f48c693c58\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" Mar 20 11:11:45 crc kubenswrapper[4846]: I0320 11:11:45.091729 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" Mar 20 11:11:45 crc kubenswrapper[4846]: I0320 11:11:45.562754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h"] Mar 20 11:11:45 crc kubenswrapper[4846]: W0320 11:11:45.582891 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5db89f9f_d51b_430f_88de_34f48c693c58.slice/crio-bf9ad2a0eeb8b2486cce6bd33ba433a86b68223ceeebe23f459f5f4e90ef49c9 WatchSource:0}: Error finding container bf9ad2a0eeb8b2486cce6bd33ba433a86b68223ceeebe23f459f5f4e90ef49c9: Status 404 returned error can't find the container with id bf9ad2a0eeb8b2486cce6bd33ba433a86b68223ceeebe23f459f5f4e90ef49c9 Mar 20 11:11:45 crc kubenswrapper[4846]: I0320 11:11:45.643568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" event={"ID":"5db89f9f-d51b-430f-88de-34f48c693c58","Type":"ContainerStarted","Data":"bf9ad2a0eeb8b2486cce6bd33ba433a86b68223ceeebe23f459f5f4e90ef49c9"} Mar 20 11:11:46 crc kubenswrapper[4846]: I0320 11:11:46.785090 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:46 crc kubenswrapper[4846]: I0320 11:11:46.785474 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:47 crc kubenswrapper[4846]: I0320 11:11:47.832709 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pxg7s" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="registry-server" probeResult="failure" output=< Mar 20 11:11:47 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:11:47 crc kubenswrapper[4846]: > Mar 20 11:11:49 crc kubenswrapper[4846]: I0320 11:11:49.671210 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" event={"ID":"5db89f9f-d51b-430f-88de-34f48c693c58","Type":"ContainerStarted","Data":"4c5c006fe43411be215768214cf9b9b4e646b9fe162ec21cb6577b3a3cde8a37"} Mar 20 11:11:49 crc kubenswrapper[4846]: I0320 11:11:49.695975 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-fcc6h" podStartSLOduration=2.518646815 podStartE2EDuration="5.695957388s" podCreationTimestamp="2026-03-20 11:11:44 +0000 UTC" firstStartedPulling="2026-03-20 11:11:45.584511058 +0000 UTC m=+876.950919294" lastFinishedPulling="2026-03-20 11:11:48.761821631 +0000 UTC m=+880.128229867" observedRunningTime="2026-03-20 11:11:49.690494964 +0000 UTC m=+881.056903220" watchObservedRunningTime="2026-03-20 11:11:49.695957388 +0000 UTC m=+881.062365624" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.446413 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.448266 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.450535 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-w5444" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.451609 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.452659 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.454513 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.474846 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.478230 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.482518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6kkm\" (UniqueName: \"kubernetes.io/projected/13c7db72-fd6b-4cdc-998a-163bec94b0f3-kube-api-access-k6kkm\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.482578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/13c7db72-fd6b-4cdc-998a-163bec94b0f3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.482608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2k5\" (UniqueName: \"kubernetes.io/projected/8bdeb909-49d8-4de7-98dc-ac1f1466dc98-kube-api-access-fm2k5\") pod \"nmstate-metrics-9b8c8685d-pzj9s\" (UID: \"8bdeb909-49d8-4de7-98dc-ac1f1466dc98\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.492422 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-cl5nt"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.493365 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.583946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-ovs-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584021 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6kkm\" (UniqueName: \"kubernetes.io/projected/13c7db72-fd6b-4cdc-998a-163bec94b0f3-kube-api-access-k6kkm\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/13c7db72-fd6b-4cdc-998a-163bec94b0f3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584475 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2k5\" (UniqueName: \"kubernetes.io/projected/8bdeb909-49d8-4de7-98dc-ac1f1466dc98-kube-api-access-fm2k5\") pod \"nmstate-metrics-9b8c8685d-pzj9s\" (UID: \"8bdeb909-49d8-4de7-98dc-ac1f1466dc98\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584547 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs2t5\" (UniqueName: \"kubernetes.io/projected/0c693da7-8768-44e0-a24a-d1f770cf0e4a-kube-api-access-fs2t5\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-dbus-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.584626 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-nmstate-lock\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.586753 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.587818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.595452 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-knrfv" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.595663 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.595741 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.596223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/13c7db72-fd6b-4cdc-998a-163bec94b0f3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.606035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2k5\" (UniqueName: \"kubernetes.io/projected/8bdeb909-49d8-4de7-98dc-ac1f1466dc98-kube-api-access-fm2k5\") pod \"nmstate-metrics-9b8c8685d-pzj9s\" (UID: \"8bdeb909-49d8-4de7-98dc-ac1f1466dc98\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.607716 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6kkm\" (UniqueName: \"kubernetes.io/projected/13c7db72-fd6b-4cdc-998a-163bec94b0f3-kube-api-access-k6kkm\") pod \"nmstate-webhook-5f558f5558-pbzwm\" (UID: \"13c7db72-fd6b-4cdc-998a-163bec94b0f3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.613426 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685234 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7c757d6d-23d9-4555-8f78-c9bd05130eae-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-nmstate-lock\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685321 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-ovs-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685339 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c757d6d-23d9-4555-8f78-c9bd05130eae-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685369 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx6g5\" (UniqueName: \"kubernetes.io/projected/7c757d6d-23d9-4555-8f78-c9bd05130eae-kube-api-access-lx6g5\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs2t5\" (UniqueName: \"kubernetes.io/projected/0c693da7-8768-44e0-a24a-d1f770cf0e4a-kube-api-access-fs2t5\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685411 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-nmstate-lock\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-dbus-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-ovs-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.685741 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0c693da7-8768-44e0-a24a-d1f770cf0e4a-dbus-socket\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.717603 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs2t5\" (UniqueName: \"kubernetes.io/projected/0c693da7-8768-44e0-a24a-d1f770cf0e4a-kube-api-access-fs2t5\") pod \"nmstate-handler-cl5nt\" (UID: \"0c693da7-8768-44e0-a24a-d1f770cf0e4a\") " pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.774555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.786623 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7c757d6d-23d9-4555-8f78-c9bd05130eae-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.786705 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c757d6d-23d9-4555-8f78-c9bd05130eae-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.786745 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx6g5\" (UniqueName: \"kubernetes.io/projected/7c757d6d-23d9-4555-8f78-c9bd05130eae-kube-api-access-lx6g5\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.787619 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7c757d6d-23d9-4555-8f78-c9bd05130eae-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.791304 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.792216 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c757d6d-23d9-4555-8f78-c9bd05130eae-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.804282 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-568d84f6d5-dmvdj"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.805791 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.813528 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.813752 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx6g5\" (UniqueName: \"kubernetes.io/projected/7c757d6d-23d9-4555-8f78-c9bd05130eae-kube-api-access-lx6g5\") pod \"nmstate-console-plugin-86f58fcf4-mpkwp\" (UID: \"7c757d6d-23d9-4555-8f78-c9bd05130eae\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.833178 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-568d84f6d5-dmvdj"] Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.887770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-oauth-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888198 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j54f2\" (UniqueName: \"kubernetes.io/projected/7fc98d21-a016-44d4-a18d-407a92b3e1d8-kube-api-access-j54f2\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888254 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888283 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-service-ca\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-trusted-ca-bundle\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.888384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-oauth-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.950330 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-service-ca\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-trusted-ca-bundle\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989271 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-oauth-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-oauth-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.989389 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j54f2\" (UniqueName: \"kubernetes.io/projected/7fc98d21-a016-44d4-a18d-407a92b3e1d8-kube-api-access-j54f2\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.990359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-service-ca\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.990511 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.991612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-trusted-ca-bundle\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.991652 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7fc98d21-a016-44d4-a18d-407a92b3e1d8-oauth-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:54 crc kubenswrapper[4846]: I0320 11:11:54.994673 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-serving-cert\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.004228 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7fc98d21-a016-44d4-a18d-407a92b3e1d8-console-oauth-config\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.020029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j54f2\" (UniqueName: \"kubernetes.io/projected/7fc98d21-a016-44d4-a18d-407a92b3e1d8-kube-api-access-j54f2\") pod \"console-568d84f6d5-dmvdj\" (UID: \"7fc98d21-a016-44d4-a18d-407a92b3e1d8\") " pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.144328 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.147605 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s"] Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.234835 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm"] Mar 20 11:11:55 crc kubenswrapper[4846]: W0320 11:11:55.245487 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13c7db72_fd6b_4cdc_998a_163bec94b0f3.slice/crio-d191d0e507135ad070c4852a57a87755e106bef66f9abb0ce445ed87be0c6620 WatchSource:0}: Error finding container d191d0e507135ad070c4852a57a87755e106bef66f9abb0ce445ed87be0c6620: Status 404 returned error can't find the container with id d191d0e507135ad070c4852a57a87755e106bef66f9abb0ce445ed87be0c6620 Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.381361 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp"] Mar 20 11:11:55 crc kubenswrapper[4846]: W0320 11:11:55.388648 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c757d6d_23d9_4555_8f78_c9bd05130eae.slice/crio-b7490ab922ab41654ba4a17d6af53e37631b3111c09065c813a42bf1733f0747 WatchSource:0}: Error finding container b7490ab922ab41654ba4a17d6af53e37631b3111c09065c813a42bf1733f0747: Status 404 returned error can't find the container with id b7490ab922ab41654ba4a17d6af53e37631b3111c09065c813a42bf1733f0747 Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.449363 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-568d84f6d5-dmvdj"] Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.719924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" event={"ID":"8bdeb909-49d8-4de7-98dc-ac1f1466dc98","Type":"ContainerStarted","Data":"d07f40922fa2b793215a4baaa4899359b4894338d69b20822f57efbb7b0e3188"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.721684 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" event={"ID":"13c7db72-fd6b-4cdc-998a-163bec94b0f3","Type":"ContainerStarted","Data":"d191d0e507135ad070c4852a57a87755e106bef66f9abb0ce445ed87be0c6620"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.723009 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cl5nt" event={"ID":"0c693da7-8768-44e0-a24a-d1f770cf0e4a","Type":"ContainerStarted","Data":"742521c5a26d1b7ebef2762d62a2924744fe09206bfe2ab06484c5fcbc73da72"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.724308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" event={"ID":"7c757d6d-23d9-4555-8f78-c9bd05130eae","Type":"ContainerStarted","Data":"b7490ab922ab41654ba4a17d6af53e37631b3111c09065c813a42bf1733f0747"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.725964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568d84f6d5-dmvdj" event={"ID":"7fc98d21-a016-44d4-a18d-407a92b3e1d8","Type":"ContainerStarted","Data":"74c2df36d6a236e5aba70aa630f4b3577910c1b1f87922908d12a6e549c885cb"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.725996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568d84f6d5-dmvdj" event={"ID":"7fc98d21-a016-44d4-a18d-407a92b3e1d8","Type":"ContainerStarted","Data":"3a09030b4f776036a8a3642a7d50da3b8db56b5cef3512b08b7fac1850d661b5"} Mar 20 11:11:55 crc kubenswrapper[4846]: I0320 11:11:55.744797 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-568d84f6d5-dmvdj" podStartSLOduration=1.744771267 podStartE2EDuration="1.744771267s" podCreationTimestamp="2026-03-20 11:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:11:55.743553879 +0000 UTC m=+887.109962125" watchObservedRunningTime="2026-03-20 11:11:55.744771267 +0000 UTC m=+887.111179503" Mar 20 11:11:56 crc kubenswrapper[4846]: I0320 11:11:56.840751 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:56 crc kubenswrapper[4846]: I0320 11:11:56.886719 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:57 crc kubenswrapper[4846]: I0320 11:11:57.075379 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:58 crc kubenswrapper[4846]: I0320 11:11:58.753191 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxg7s" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="registry-server" containerID="cri-o://b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8" gracePeriod=2 Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.099292 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.152163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content\") pod \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.152320 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities\") pod \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.153243 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities" (OuterVolumeSpecName: "utilities") pod "a5f0029a-cf4c-4abf-b800-e617cba4ac3b" (UID: "a5f0029a-cf4c-4abf-b800-e617cba4ac3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.253674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp9vt\" (UniqueName: \"kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt\") pod \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\" (UID: \"a5f0029a-cf4c-4abf-b800-e617cba4ac3b\") " Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.253881 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.263578 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt" (OuterVolumeSpecName: "kube-api-access-zp9vt") pod "a5f0029a-cf4c-4abf-b800-e617cba4ac3b" (UID: "a5f0029a-cf4c-4abf-b800-e617cba4ac3b"). InnerVolumeSpecName "kube-api-access-zp9vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.294304 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5f0029a-cf4c-4abf-b800-e617cba4ac3b" (UID: "a5f0029a-cf4c-4abf-b800-e617cba4ac3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.360234 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp9vt\" (UniqueName: \"kubernetes.io/projected/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-kube-api-access-zp9vt\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.360304 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5f0029a-cf4c-4abf-b800-e617cba4ac3b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.769917 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" event={"ID":"13c7db72-fd6b-4cdc-998a-163bec94b0f3","Type":"ContainerStarted","Data":"50555283e426cbbe0cf6bb923c8982ac7b4eb0aeb56b405af7ebe779127c5944"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.770421 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.776068 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerID="b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8" exitCode=0 Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.776156 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxg7s" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.776158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerDied","Data":"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.776313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxg7s" event={"ID":"a5f0029a-cf4c-4abf-b800-e617cba4ac3b","Type":"ContainerDied","Data":"aa2264ae0db4071e9810408bef99b5514bed418d6af48d8388b36934a8b87c60"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.776338 4846 scope.go:117] "RemoveContainer" containerID="b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.781165 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cl5nt" event={"ID":"0c693da7-8768-44e0-a24a-d1f770cf0e4a","Type":"ContainerStarted","Data":"b61d0193f14a86a8d83198f706476bb41e2bd93ec30246ceed1e53d9fba07be7"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.781294 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.783716 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" event={"ID":"7c757d6d-23d9-4555-8f78-c9bd05130eae","Type":"ContainerStarted","Data":"c11d4b70ca4ed078875f1d81925990d61fc1521f84519f95a2db58790f857930"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.791252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" event={"ID":"8bdeb909-49d8-4de7-98dc-ac1f1466dc98","Type":"ContainerStarted","Data":"2a7a4a094ce07eb1f1ef31c049befb7913f11c8f17e23846c0b8b29e0c8c9f16"} Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.794847 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" podStartSLOduration=2.292661361 podStartE2EDuration="5.794823084s" podCreationTimestamp="2026-03-20 11:11:54 +0000 UTC" firstStartedPulling="2026-03-20 11:11:55.257564209 +0000 UTC m=+886.623972445" lastFinishedPulling="2026-03-20 11:11:58.759725932 +0000 UTC m=+890.126134168" observedRunningTime="2026-03-20 11:11:59.788741074 +0000 UTC m=+891.155149310" watchObservedRunningTime="2026-03-20 11:11:59.794823084 +0000 UTC m=+891.161231320" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.822807 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-cl5nt" podStartSLOduration=1.991071606 podStartE2EDuration="5.822780967s" podCreationTimestamp="2026-03-20 11:11:54 +0000 UTC" firstStartedPulling="2026-03-20 11:11:54.880411845 +0000 UTC m=+886.246820081" lastFinishedPulling="2026-03-20 11:11:58.712121206 +0000 UTC m=+890.078529442" observedRunningTime="2026-03-20 11:11:59.822164374 +0000 UTC m=+891.188572610" watchObservedRunningTime="2026-03-20 11:11:59.822780967 +0000 UTC m=+891.189189203" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.830829 4846 scope.go:117] "RemoveContainer" containerID="78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.842775 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-mpkwp" podStartSLOduration=2.519943432 podStartE2EDuration="5.842756357s" podCreationTimestamp="2026-03-20 11:11:54 +0000 UTC" firstStartedPulling="2026-03-20 11:11:55.390368196 +0000 UTC m=+886.756776432" lastFinishedPulling="2026-03-20 11:11:58.713181121 +0000 UTC m=+890.079589357" observedRunningTime="2026-03-20 11:11:59.839657406 +0000 UTC m=+891.206065652" watchObservedRunningTime="2026-03-20 11:11:59.842756357 +0000 UTC m=+891.209164593" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.866092 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.872165 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxg7s"] Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.884365 4846 scope.go:117] "RemoveContainer" containerID="e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.915738 4846 scope.go:117] "RemoveContainer" containerID="b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8" Mar 20 11:11:59 crc kubenswrapper[4846]: E0320 11:11:59.916121 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8\": container with ID starting with b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8 not found: ID does not exist" containerID="b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.916161 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8"} err="failed to get container status \"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8\": rpc error: code = NotFound desc = could not find container \"b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8\": container with ID starting with b2c1ea5554b15487ea99da3e1469bbe39586e14e1ea847a48081edf5be980df8 not found: ID does not exist" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.916224 4846 scope.go:117] "RemoveContainer" containerID="78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54" Mar 20 11:11:59 crc kubenswrapper[4846]: E0320 11:11:59.916528 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54\": container with ID starting with 78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54 not found: ID does not exist" containerID="78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.916551 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54"} err="failed to get container status \"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54\": rpc error: code = NotFound desc = could not find container \"78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54\": container with ID starting with 78c0f201c3349e3da000b3def94d893adea078cf5d0e94ce115690d0dabc0b54 not found: ID does not exist" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.916565 4846 scope.go:117] "RemoveContainer" containerID="e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99" Mar 20 11:11:59 crc kubenswrapper[4846]: E0320 11:11:59.917017 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99\": container with ID starting with e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99 not found: ID does not exist" containerID="e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99" Mar 20 11:11:59 crc kubenswrapper[4846]: I0320 11:11:59.917093 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99"} err="failed to get container status \"e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99\": rpc error: code = NotFound desc = could not find container \"e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99\": container with ID starting with e5cf6d3049479306b41153360ecf9af69b0e494bfe9c77917e25e28bbc1c6b99 not found: ID does not exist" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.133819 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566752-qrzct"] Mar 20 11:12:00 crc kubenswrapper[4846]: E0320 11:12:00.137631 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="extract-utilities" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.137694 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="extract-utilities" Mar 20 11:12:00 crc kubenswrapper[4846]: E0320 11:12:00.137707 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="extract-content" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.137735 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="extract-content" Mar 20 11:12:00 crc kubenswrapper[4846]: E0320 11:12:00.137755 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="registry-server" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.137767 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="registry-server" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.137988 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" containerName="registry-server" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.138686 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.141754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566752-qrzct"] Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.141802 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.142020 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.142256 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.172237 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9ll4\" (UniqueName: \"kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4\") pod \"auto-csr-approver-29566752-qrzct\" (UID: \"cd72fbad-b2b4-44b0-8de2-b6209cb883a5\") " pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.273428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9ll4\" (UniqueName: \"kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4\") pod \"auto-csr-approver-29566752-qrzct\" (UID: \"cd72fbad-b2b4-44b0-8de2-b6209cb883a5\") " pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.298470 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9ll4\" (UniqueName: \"kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4\") pod \"auto-csr-approver-29566752-qrzct\" (UID: \"cd72fbad-b2b4-44b0-8de2-b6209cb883a5\") " pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.463333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.673278 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566752-qrzct"] Mar 20 11:12:00 crc kubenswrapper[4846]: I0320 11:12:00.805098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566752-qrzct" event={"ID":"cd72fbad-b2b4-44b0-8de2-b6209cb883a5","Type":"ContainerStarted","Data":"8344c6af5e20dcceca760e4733d9b4411f4d05ce020a6604628941fa7a8e71db"} Mar 20 11:12:01 crc kubenswrapper[4846]: I0320 11:12:01.332367 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f0029a-cf4c-4abf-b800-e617cba4ac3b" path="/var/lib/kubelet/pods/a5f0029a-cf4c-4abf-b800-e617cba4ac3b/volumes" Mar 20 11:12:01 crc kubenswrapper[4846]: I0320 11:12:01.812401 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" event={"ID":"8bdeb909-49d8-4de7-98dc-ac1f1466dc98","Type":"ContainerStarted","Data":"bd2784c4e5a7f316dcb11eab542858867ac69fba3e45b8c4d08d8a8e31a3e4b8"} Mar 20 11:12:01 crc kubenswrapper[4846]: I0320 11:12:01.836214 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-pzj9s" podStartSLOduration=1.571737743 podStartE2EDuration="7.836183824s" podCreationTimestamp="2026-03-20 11:11:54 +0000 UTC" firstStartedPulling="2026-03-20 11:11:55.184361953 +0000 UTC m=+886.550770189" lastFinishedPulling="2026-03-20 11:12:01.448808034 +0000 UTC m=+892.815216270" observedRunningTime="2026-03-20 11:12:01.829752886 +0000 UTC m=+893.196161112" watchObservedRunningTime="2026-03-20 11:12:01.836183824 +0000 UTC m=+893.202592060" Mar 20 11:12:02 crc kubenswrapper[4846]: I0320 11:12:02.825516 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd72fbad-b2b4-44b0-8de2-b6209cb883a5" containerID="f97f4aac2dc141a4c63059d1b591089ffec6a0503f2c63998c10ca35e916f80d" exitCode=0 Mar 20 11:12:02 crc kubenswrapper[4846]: I0320 11:12:02.825573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566752-qrzct" event={"ID":"cd72fbad-b2b4-44b0-8de2-b6209cb883a5","Type":"ContainerDied","Data":"f97f4aac2dc141a4c63059d1b591089ffec6a0503f2c63998c10ca35e916f80d"} Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.077726 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.229731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9ll4\" (UniqueName: \"kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4\") pod \"cd72fbad-b2b4-44b0-8de2-b6209cb883a5\" (UID: \"cd72fbad-b2b4-44b0-8de2-b6209cb883a5\") " Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.238688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4" (OuterVolumeSpecName: "kube-api-access-x9ll4") pod "cd72fbad-b2b4-44b0-8de2-b6209cb883a5" (UID: "cd72fbad-b2b4-44b0-8de2-b6209cb883a5"). InnerVolumeSpecName "kube-api-access-x9ll4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.331695 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9ll4\" (UniqueName: \"kubernetes.io/projected/cd72fbad-b2b4-44b0-8de2-b6209cb883a5-kube-api-access-x9ll4\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.837472 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-cl5nt" Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.840726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566752-qrzct" event={"ID":"cd72fbad-b2b4-44b0-8de2-b6209cb883a5","Type":"ContainerDied","Data":"8344c6af5e20dcceca760e4733d9b4411f4d05ce020a6604628941fa7a8e71db"} Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.840765 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8344c6af5e20dcceca760e4733d9b4411f4d05ce020a6604628941fa7a8e71db" Mar 20 11:12:04 crc kubenswrapper[4846]: I0320 11:12:04.840821 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566752-qrzct" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.144544 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566746-fkvh8"] Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.145427 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.145477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.149667 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566746-fkvh8"] Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.152583 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.331464 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3" path="/var/lib/kubelet/pods/5349cbfc-22ff-4a9d-8ed8-f4c17fae41f3/volumes" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.851404 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-568d84f6d5-dmvdj" Mar 20 11:12:05 crc kubenswrapper[4846]: I0320 11:12:05.903134 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:12:13 crc kubenswrapper[4846]: I0320 11:12:13.732462 4846 scope.go:117] "RemoveContainer" containerID="a8a327098bfd4307ec579af79ebc7d1cc71cf40b1512bdcfc9616765affb0a9f" Mar 20 11:12:14 crc kubenswrapper[4846]: I0320 11:12:14.799200 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-pbzwm" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.299091 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d"] Mar 20 11:12:27 crc kubenswrapper[4846]: E0320 11:12:27.300113 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd72fbad-b2b4-44b0-8de2-b6209cb883a5" containerName="oc" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.300131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd72fbad-b2b4-44b0-8de2-b6209cb883a5" containerName="oc" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.300262 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd72fbad-b2b4-44b0-8de2-b6209cb883a5" containerName="oc" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.301565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.303995 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.310564 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d"] Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.447669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.447835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.447865 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5whh\" (UniqueName: \"kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.549237 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.549305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.549329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5whh\" (UniqueName: \"kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.550278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.550416 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.577456 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5whh\" (UniqueName: \"kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:27 crc kubenswrapper[4846]: I0320 11:12:27.628437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:28 crc kubenswrapper[4846]: I0320 11:12:28.051999 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d"] Mar 20 11:12:28 crc kubenswrapper[4846]: I0320 11:12:28.991876 4846 generic.go:334] "Generic (PLEG): container finished" podID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerID="6283bd5cfaf2eb5264ab759b4235230ce571ec14eba7a0faf64bf6b94775fff1" exitCode=0 Mar 20 11:12:28 crc kubenswrapper[4846]: I0320 11:12:28.991990 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" event={"ID":"a07c8578-7175-4eb7-9c40-416863db9cb2","Type":"ContainerDied","Data":"6283bd5cfaf2eb5264ab759b4235230ce571ec14eba7a0faf64bf6b94775fff1"} Mar 20 11:12:28 crc kubenswrapper[4846]: I0320 11:12:28.992207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" event={"ID":"a07c8578-7175-4eb7-9c40-416863db9cb2","Type":"ContainerStarted","Data":"2f505eb9aee5d0a9e42dc6374c0469185c35e56f8819cbe32bc64f8803e9ac93"} Mar 20 11:12:30 crc kubenswrapper[4846]: I0320 11:12:30.961070 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-bwgwb" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" containerID="cri-o://24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989" gracePeriod=15 Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.007066 4846 generic.go:334] "Generic (PLEG): container finished" podID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerID="f8997cb09897bc54835f6fc5649a8786967acefed52fcc4f6da427a8fccadfc7" exitCode=0 Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.007129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" event={"ID":"a07c8578-7175-4eb7-9c40-416863db9cb2","Type":"ContainerDied","Data":"f8997cb09897bc54835f6fc5649a8786967acefed52fcc4f6da427a8fccadfc7"} Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.369150 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bwgwb_7b910eda-ae6a-4e2e-9ce2-9a139756b29f/console/0.log" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.369513 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504637 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504710 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504737 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.504827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpl2s\" (UniqueName: \"kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s\") pod \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\" (UID: \"7b910eda-ae6a-4e2e-9ce2-9a139756b29f\") " Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.505971 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.505981 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca" (OuterVolumeSpecName: "service-ca") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.506069 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config" (OuterVolumeSpecName: "console-config") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.506218 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.511591 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.511653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s" (OuterVolumeSpecName: "kube-api-access-bpl2s") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "kube-api-access-bpl2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.511847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7b910eda-ae6a-4e2e-9ce2-9a139756b29f" (UID: "7b910eda-ae6a-4e2e-9ce2-9a139756b29f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.606906 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607261 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607270 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607281 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607289 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpl2s\" (UniqueName: \"kubernetes.io/projected/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-kube-api-access-bpl2s\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607302 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:31 crc kubenswrapper[4846]: I0320 11:12:31.607310 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b910eda-ae6a-4e2e-9ce2-9a139756b29f-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.016211 4846 generic.go:334] "Generic (PLEG): container finished" podID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerID="a539ea14df5a4f107f60f37e3c1e00ade01febd8d1c15b97ea95cfc466611389" exitCode=0 Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.016297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" event={"ID":"a07c8578-7175-4eb7-9c40-416863db9cb2","Type":"ContainerDied","Data":"a539ea14df5a4f107f60f37e3c1e00ade01febd8d1c15b97ea95cfc466611389"} Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019493 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bwgwb_7b910eda-ae6a-4e2e-9ce2-9a139756b29f/console/0.log" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019528 4846 generic.go:334] "Generic (PLEG): container finished" podID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerID="24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989" exitCode=2 Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bwgwb" event={"ID":"7b910eda-ae6a-4e2e-9ce2-9a139756b29f","Type":"ContainerDied","Data":"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989"} Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bwgwb" event={"ID":"7b910eda-ae6a-4e2e-9ce2-9a139756b29f","Type":"ContainerDied","Data":"5058fafd6d57803c4c4b220f604bfd3b42e8d1ba0ed2723b35979c5f18e4f2e3"} Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019588 4846 scope.go:117] "RemoveContainer" containerID="24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.019722 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bwgwb" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.044396 4846 scope.go:117] "RemoveContainer" containerID="24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989" Mar 20 11:12:32 crc kubenswrapper[4846]: E0320 11:12:32.045942 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989\": container with ID starting with 24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989 not found: ID does not exist" containerID="24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.046021 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989"} err="failed to get container status \"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989\": rpc error: code = NotFound desc = could not find container \"24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989\": container with ID starting with 24e9d2fd66f5dc99d9fa2d8855d27402734008be9d3c327adbd9348e95052989 not found: ID does not exist" Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.064499 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:12:32 crc kubenswrapper[4846]: I0320 11:12:32.068152 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-bwgwb"] Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.247160 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.330830 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" path="/var/lib/kubelet/pods/7b910eda-ae6a-4e2e-9ce2-9a139756b29f/volumes" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.341299 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle\") pod \"a07c8578-7175-4eb7-9c40-416863db9cb2\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.341388 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5whh\" (UniqueName: \"kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh\") pod \"a07c8578-7175-4eb7-9c40-416863db9cb2\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.341450 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util\") pod \"a07c8578-7175-4eb7-9c40-416863db9cb2\" (UID: \"a07c8578-7175-4eb7-9c40-416863db9cb2\") " Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.342470 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle" (OuterVolumeSpecName: "bundle") pod "a07c8578-7175-4eb7-9c40-416863db9cb2" (UID: "a07c8578-7175-4eb7-9c40-416863db9cb2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.347267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh" (OuterVolumeSpecName: "kube-api-access-k5whh") pod "a07c8578-7175-4eb7-9c40-416863db9cb2" (UID: "a07c8578-7175-4eb7-9c40-416863db9cb2"). InnerVolumeSpecName "kube-api-access-k5whh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.360060 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util" (OuterVolumeSpecName: "util") pod "a07c8578-7175-4eb7-9c40-416863db9cb2" (UID: "a07c8578-7175-4eb7-9c40-416863db9cb2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.443491 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-util\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.443537 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a07c8578-7175-4eb7-9c40-416863db9cb2-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:33 crc kubenswrapper[4846]: I0320 11:12:33.443547 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5whh\" (UniqueName: \"kubernetes.io/projected/a07c8578-7175-4eb7-9c40-416863db9cb2-kube-api-access-k5whh\") on node \"crc\" DevicePath \"\"" Mar 20 11:12:34 crc kubenswrapper[4846]: I0320 11:12:34.036312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" event={"ID":"a07c8578-7175-4eb7-9c40-416863db9cb2","Type":"ContainerDied","Data":"2f505eb9aee5d0a9e42dc6374c0469185c35e56f8819cbe32bc64f8803e9ac93"} Mar 20 11:12:34 crc kubenswrapper[4846]: I0320 11:12:34.036367 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f505eb9aee5d0a9e42dc6374c0469185c35e56f8819cbe32bc64f8803e9ac93" Mar 20 11:12:34 crc kubenswrapper[4846]: I0320 11:12:34.036371 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.473859 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj"] Mar 20 11:12:42 crc kubenswrapper[4846]: E0320 11:12:42.474933 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.474952 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" Mar 20 11:12:42 crc kubenswrapper[4846]: E0320 11:12:42.474965 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="util" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.474973 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="util" Mar 20 11:12:42 crc kubenswrapper[4846]: E0320 11:12:42.474987 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="extract" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.474994 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="extract" Mar 20 11:12:42 crc kubenswrapper[4846]: E0320 11:12:42.475011 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="pull" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.475019 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="pull" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.475144 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a07c8578-7175-4eb7-9c40-416863db9cb2" containerName="extract" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.475160 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b910eda-ae6a-4e2e-9ce2-9a139756b29f" containerName="console" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.475770 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.477606 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.478346 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.479049 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.479079 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jw5s9" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.480432 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.491205 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj"] Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.587558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw6bw\" (UniqueName: \"kubernetes.io/projected/78d82817-160a-4eb0-96ba-da47a62bd61c-kube-api-access-zw6bw\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.587693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-apiservice-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.587735 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-webhook-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.688555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw6bw\" (UniqueName: \"kubernetes.io/projected/78d82817-160a-4eb0-96ba-da47a62bd61c-kube-api-access-zw6bw\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.688620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-apiservice-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.688642 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-webhook-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.696143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-webhook-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.696549 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/78d82817-160a-4eb0-96ba-da47a62bd61c-apiservice-cert\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.710685 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw6bw\" (UniqueName: \"kubernetes.io/projected/78d82817-160a-4eb0-96ba-da47a62bd61c-kube-api-access-zw6bw\") pod \"metallb-operator-controller-manager-66844494d6-8t8lj\" (UID: \"78d82817-160a-4eb0-96ba-da47a62bd61c\") " pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.795216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.836157 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq"] Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.837007 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.850280 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.850288 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fvmnx" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.850450 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.867398 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq"] Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.893606 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rfh9\" (UniqueName: \"kubernetes.io/projected/5a3417e0-ef3f-461d-b798-caab9527c70b-kube-api-access-6rfh9\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.893711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-apiservice-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.893740 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-webhook-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.994972 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rfh9\" (UniqueName: \"kubernetes.io/projected/5a3417e0-ef3f-461d-b798-caab9527c70b-kube-api-access-6rfh9\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.995164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-apiservice-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:42 crc kubenswrapper[4846]: I0320 11:12:42.995213 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-webhook-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.004153 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-webhook-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.020331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rfh9\" (UniqueName: \"kubernetes.io/projected/5a3417e0-ef3f-461d-b798-caab9527c70b-kube-api-access-6rfh9\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.022174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a3417e0-ef3f-461d-b798-caab9527c70b-apiservice-cert\") pod \"metallb-operator-webhook-server-cd87fcd9-tfmhq\" (UID: \"5a3417e0-ef3f-461d-b798-caab9527c70b\") " pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.177679 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.360659 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj"] Mar 20 11:12:43 crc kubenswrapper[4846]: W0320 11:12:43.471970 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a3417e0_ef3f_461d_b798_caab9527c70b.slice/crio-08b8d36b8c5c90b728e9b0ab6597c580d54f62353c5ce80b40b326db6836412c WatchSource:0}: Error finding container 08b8d36b8c5c90b728e9b0ab6597c580d54f62353c5ce80b40b326db6836412c: Status 404 returned error can't find the container with id 08b8d36b8c5c90b728e9b0ab6597c580d54f62353c5ce80b40b326db6836412c Mar 20 11:12:43 crc kubenswrapper[4846]: I0320 11:12:43.472190 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq"] Mar 20 11:12:44 crc kubenswrapper[4846]: I0320 11:12:44.094379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" event={"ID":"5a3417e0-ef3f-461d-b798-caab9527c70b","Type":"ContainerStarted","Data":"08b8d36b8c5c90b728e9b0ab6597c580d54f62353c5ce80b40b326db6836412c"} Mar 20 11:12:44 crc kubenswrapper[4846]: I0320 11:12:44.095597 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" event={"ID":"78d82817-160a-4eb0-96ba-da47a62bd61c","Type":"ContainerStarted","Data":"868ba800e760ecb9730e1cec02069b3e63030516280835c6b30ab8ff50a60d94"} Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.145314 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" event={"ID":"5a3417e0-ef3f-461d-b798-caab9527c70b","Type":"ContainerStarted","Data":"3b569ace902e28da8fbf0af30e5522f16c54fbcfe295320c332b7622aa34d772"} Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.145781 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.149463 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" event={"ID":"78d82817-160a-4eb0-96ba-da47a62bd61c","Type":"ContainerStarted","Data":"8582ec55e0d8532a051e48f15097297f8aaf4cb0f916246516212bc0fa1e7c2c"} Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.149614 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.171469 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" podStartSLOduration=2.282876083 podStartE2EDuration="8.171442508s" podCreationTimestamp="2026-03-20 11:12:42 +0000 UTC" firstStartedPulling="2026-03-20 11:12:43.478314284 +0000 UTC m=+934.844722530" lastFinishedPulling="2026-03-20 11:12:49.366880719 +0000 UTC m=+940.733288955" observedRunningTime="2026-03-20 11:12:50.166534251 +0000 UTC m=+941.532942507" watchObservedRunningTime="2026-03-20 11:12:50.171442508 +0000 UTC m=+941.537850744" Mar 20 11:12:50 crc kubenswrapper[4846]: I0320 11:12:50.197261 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" podStartSLOduration=2.222601048 podStartE2EDuration="8.197227621s" podCreationTimestamp="2026-03-20 11:12:42 +0000 UTC" firstStartedPulling="2026-03-20 11:12:43.37223183 +0000 UTC m=+934.738640066" lastFinishedPulling="2026-03-20 11:12:49.346858403 +0000 UTC m=+940.713266639" observedRunningTime="2026-03-20 11:12:50.19634029 +0000 UTC m=+941.562748536" watchObservedRunningTime="2026-03-20 11:12:50.197227621 +0000 UTC m=+941.563635857" Mar 20 11:13:03 crc kubenswrapper[4846]: I0320 11:13:03.186698 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-cd87fcd9-tfmhq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.267908 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.272265 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.335864 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.389794 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5rbl\" (UniqueName: \"kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.389861 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.389921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.491827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5rbl\" (UniqueName: \"kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.491942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.492104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.492811 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.492870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.517663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5rbl\" (UniqueName: \"kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl\") pod \"redhat-marketplace-mncwq\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:20 crc kubenswrapper[4846]: I0320 11:13:20.598316 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:21 crc kubenswrapper[4846]: I0320 11:13:21.132560 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:21 crc kubenswrapper[4846]: I0320 11:13:21.357711 4846 generic.go:334] "Generic (PLEG): container finished" podID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerID="370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9" exitCode=0 Mar 20 11:13:21 crc kubenswrapper[4846]: I0320 11:13:21.357781 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerDied","Data":"370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9"} Mar 20 11:13:21 crc kubenswrapper[4846]: I0320 11:13:21.358237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerStarted","Data":"29ce9acf7d9b34988b6e12243dae9f32b125b0634a972142f6aa3020b6074fa4"} Mar 20 11:13:22 crc kubenswrapper[4846]: I0320 11:13:22.378272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerStarted","Data":"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65"} Mar 20 11:13:22 crc kubenswrapper[4846]: I0320 11:13:22.801848 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66844494d6-8t8lj" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.389455 4846 generic.go:334] "Generic (PLEG): container finished" podID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerID="dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65" exitCode=0 Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.389534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerDied","Data":"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65"} Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.536101 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.537406 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.540616 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-d554q"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.543186 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.544943 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.545366 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.545599 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.545948 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-97c2c" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.557600 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644746 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rgdr\" (UniqueName: \"kubernetes.io/projected/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-kube-api-access-4rgdr\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644821 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644849 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics-certs\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644944 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tzrl\" (UniqueName: \"kubernetes.io/projected/98b51565-aa74-437c-9d42-a93f2b6b52de-kube-api-access-5tzrl\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.644994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-sockets\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.645022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-conf\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.645064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-reloader\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.645090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-startup\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.652632 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-jjwfh"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.654031 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.656593 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.656795 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-vmhqp" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.659049 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.662040 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.690447 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-hvrww"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.691517 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.693733 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.706643 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-hvrww"] Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tzrl\" (UniqueName: \"kubernetes.io/projected/98b51565-aa74-437c-9d42-a93f2b6b52de-kube-api-access-5tzrl\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746864 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-sockets\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746908 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-conf\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-reloader\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.746969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-startup\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rgdr\" (UniqueName: \"kubernetes.io/projected/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-kube-api-access-4rgdr\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747037 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics-certs\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-sockets\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747623 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-conf\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: E0320 11:13:23.747700 4846 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 20 11:13:23 crc kubenswrapper[4846]: E0320 11:13:23.747757 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert podName:98b51565-aa74-437c-9d42-a93f2b6b52de nodeName:}" failed. No retries permitted until 2026-03-20 11:13:24.247736932 +0000 UTC m=+975.614145168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert") pod "frr-k8s-webhook-server-bcc4b6f68-96pp8" (UID: "98b51565-aa74-437c-9d42-a93f2b6b52de") : secret "frr-k8s-webhook-server-cert" not found Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.747880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-reloader\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.748314 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-frr-startup\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.754545 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-metrics-certs\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.765763 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rgdr\" (UniqueName: \"kubernetes.io/projected/e6c14c61-393f-4234-9d0f-bcc3cf725c6e-kube-api-access-4rgdr\") pod \"frr-k8s-d554q\" (UID: \"e6c14c61-393f-4234-9d0f-bcc3cf725c6e\") " pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.766836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tzrl\" (UniqueName: \"kubernetes.io/projected/98b51565-aa74-437c-9d42-a93f2b6b52de-kube-api-access-5tzrl\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.847919 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-cert\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.847977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-metrics-certs\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.848023 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pvmd\" (UniqueName: \"kubernetes.io/projected/bc262413-3fc0-4204-a662-b8e3c1a9eb15-kube-api-access-8pvmd\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.848049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82qxq\" (UniqueName: \"kubernetes.io/projected/08211b7a-f5ad-48aa-85ac-be288b9e3e40-kube-api-access-82qxq\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.848072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.848112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metrics-certs\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.848191 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metallb-excludel2\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.871255 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.950556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metallb-excludel2\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-cert\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951260 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-metrics-certs\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951358 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pvmd\" (UniqueName: \"kubernetes.io/projected/bc262413-3fc0-4204-a662-b8e3c1a9eb15-kube-api-access-8pvmd\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82qxq\" (UniqueName: \"kubernetes.io/projected/08211b7a-f5ad-48aa-85ac-be288b9e3e40-kube-api-access-82qxq\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951588 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metrics-certs\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.951764 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metallb-excludel2\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: E0320 11:13:23.953114 4846 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 11:13:23 crc kubenswrapper[4846]: E0320 11:13:23.953205 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist podName:08211b7a-f5ad-48aa-85ac-be288b9e3e40 nodeName:}" failed. No retries permitted until 2026-03-20 11:13:24.453181178 +0000 UTC m=+975.819589574 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist") pod "speaker-jjwfh" (UID: "08211b7a-f5ad-48aa-85ac-be288b9e3e40") : secret "metallb-memberlist" not found Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.957554 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-metrics-certs\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.958076 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.962657 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-metrics-certs\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.966790 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bc262413-3fc0-4204-a662-b8e3c1a9eb15-cert\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.974583 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pvmd\" (UniqueName: \"kubernetes.io/projected/bc262413-3fc0-4204-a662-b8e3c1a9eb15-kube-api-access-8pvmd\") pod \"controller-7bb4cc7c98-hvrww\" (UID: \"bc262413-3fc0-4204-a662-b8e3c1a9eb15\") " pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:23 crc kubenswrapper[4846]: I0320 11:13:23.978133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82qxq\" (UniqueName: \"kubernetes.io/projected/08211b7a-f5ad-48aa-85ac-be288b9e3e40-kube-api-access-82qxq\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.006746 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.241505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-hvrww"] Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.258089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.264259 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98b51565-aa74-437c-9d42-a93f2b6b52de-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-96pp8\" (UID: \"98b51565-aa74-437c-9d42-a93f2b6b52de\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.396812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"2632833088cded0e004d46ec7f5386259aaa4f8f673c70d142cafb637dae5525"} Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.398547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-hvrww" event={"ID":"bc262413-3fc0-4204-a662-b8e3c1a9eb15","Type":"ContainerStarted","Data":"9a29b64883cff35f02b6465e347442c35d119792fa596ad0020540166592be49"} Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.398606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-hvrww" event={"ID":"bc262413-3fc0-4204-a662-b8e3c1a9eb15","Type":"ContainerStarted","Data":"b92f9a50894fbb21ea1e4bf8f77b8e7ee42ae79096958e781dd03eaf56ed0e3e"} Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.400812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerStarted","Data":"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b"} Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.430357 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mncwq" podStartSLOduration=1.881726783 podStartE2EDuration="4.430328439s" podCreationTimestamp="2026-03-20 11:13:20 +0000 UTC" firstStartedPulling="2026-03-20 11:13:21.35948568 +0000 UTC m=+972.725893916" lastFinishedPulling="2026-03-20 11:13:23.908087336 +0000 UTC m=+975.274495572" observedRunningTime="2026-03-20 11:13:24.427613524 +0000 UTC m=+975.794021760" watchObservedRunningTime="2026-03-20 11:13:24.430328439 +0000 UTC m=+975.796736685" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.458812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.461029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.465599 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/08211b7a-f5ad-48aa-85ac-be288b9e3e40-memberlist\") pod \"speaker-jjwfh\" (UID: \"08211b7a-f5ad-48aa-85ac-be288b9e3e40\") " pod="metallb-system/speaker-jjwfh" Mar 20 11:13:24 crc kubenswrapper[4846]: I0320 11:13:24.571530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jjwfh" Mar 20 11:13:24 crc kubenswrapper[4846]: W0320 11:13:24.618359 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08211b7a_f5ad_48aa_85ac_be288b9e3e40.slice/crio-353cc44f81d78292399da53c6fbb145283cf28e93e79ffd896a7c76633738386 WatchSource:0}: Error finding container 353cc44f81d78292399da53c6fbb145283cf28e93e79ffd896a7c76633738386: Status 404 returned error can't find the container with id 353cc44f81d78292399da53c6fbb145283cf28e93e79ffd896a7c76633738386 Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.005286 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8"] Mar 20 11:13:25 crc kubenswrapper[4846]: W0320 11:13:25.027760 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98b51565_aa74_437c_9d42_a93f2b6b52de.slice/crio-d80070456927574442c67f618502d51da70f08122a98c4997f0a8436e7fb0b8c WatchSource:0}: Error finding container d80070456927574442c67f618502d51da70f08122a98c4997f0a8436e7fb0b8c: Status 404 returned error can't find the container with id d80070456927574442c67f618502d51da70f08122a98c4997f0a8436e7fb0b8c Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.407070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" event={"ID":"98b51565-aa74-437c-9d42-a93f2b6b52de","Type":"ContainerStarted","Data":"d80070456927574442c67f618502d51da70f08122a98c4997f0a8436e7fb0b8c"} Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.409879 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jjwfh" event={"ID":"08211b7a-f5ad-48aa-85ac-be288b9e3e40","Type":"ContainerStarted","Data":"78bc397a9d15a06768e7133c7dd4c98da289cd67810adab9ddc522e28ef143be"} Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.409919 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jjwfh" event={"ID":"08211b7a-f5ad-48aa-85ac-be288b9e3e40","Type":"ContainerStarted","Data":"4fbf67987205e0ed534a791b6656bf3781607b6de3bbaaf93a5b44d983f5db1b"} Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.409930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jjwfh" event={"ID":"08211b7a-f5ad-48aa-85ac-be288b9e3e40","Type":"ContainerStarted","Data":"353cc44f81d78292399da53c6fbb145283cf28e93e79ffd896a7c76633738386"} Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.410085 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-jjwfh" Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.412000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-hvrww" event={"ID":"bc262413-3fc0-4204-a662-b8e3c1a9eb15","Type":"ContainerStarted","Data":"3a4024c90f90aacc01c679d16eeec98181f8f42249350100fc151e5fce74d649"} Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.433681 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-jjwfh" podStartSLOduration=2.433658746 podStartE2EDuration="2.433658746s" podCreationTimestamp="2026-03-20 11:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:13:25.428945894 +0000 UTC m=+976.795354150" watchObservedRunningTime="2026-03-20 11:13:25.433658746 +0000 UTC m=+976.800066982" Mar 20 11:13:25 crc kubenswrapper[4846]: I0320 11:13:25.454234 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-hvrww" podStartSLOduration=2.454210085 podStartE2EDuration="2.454210085s" podCreationTimestamp="2026-03-20 11:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:13:25.446545503 +0000 UTC m=+976.812953739" watchObservedRunningTime="2026-03-20 11:13:25.454210085 +0000 UTC m=+976.820618321" Mar 20 11:13:26 crc kubenswrapper[4846]: I0320 11:13:26.427643 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:30 crc kubenswrapper[4846]: I0320 11:13:30.598528 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:30 crc kubenswrapper[4846]: I0320 11:13:30.598954 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:30 crc kubenswrapper[4846]: I0320 11:13:30.699443 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:31 crc kubenswrapper[4846]: I0320 11:13:31.571634 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:31 crc kubenswrapper[4846]: I0320 11:13:31.625943 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:33 crc kubenswrapper[4846]: I0320 11:13:33.506816 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mncwq" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="registry-server" containerID="cri-o://f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b" gracePeriod=2 Mar 20 11:13:33 crc kubenswrapper[4846]: I0320 11:13:33.889732 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.013789 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-hvrww" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.038502 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities\") pod \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.038616 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content\") pod \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.038690 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5rbl\" (UniqueName: \"kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl\") pod \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\" (UID: \"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9\") " Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.039825 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities" (OuterVolumeSpecName: "utilities") pod "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" (UID: "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.047124 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl" (OuterVolumeSpecName: "kube-api-access-m5rbl") pod "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" (UID: "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9"). InnerVolumeSpecName "kube-api-access-m5rbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.069284 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" (UID: "587c0cfb-32b5-4051-9ba6-a8dc13f5acf9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.140067 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.140115 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.140133 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5rbl\" (UniqueName: \"kubernetes.io/projected/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9-kube-api-access-m5rbl\") on node \"crc\" DevicePath \"\"" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.514327 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c14c61-393f-4234-9d0f-bcc3cf725c6e" containerID="c7fbbf811e9229a83c4f888c48c184ab3f7b78b5e1174701d1effa06c36c4ac8" exitCode=0 Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.514405 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerDied","Data":"c7fbbf811e9229a83c4f888c48c184ab3f7b78b5e1174701d1effa06c36c4ac8"} Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.517713 4846 generic.go:334] "Generic (PLEG): container finished" podID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerID="f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b" exitCode=0 Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.517764 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mncwq" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.517792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerDied","Data":"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b"} Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.517828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mncwq" event={"ID":"587c0cfb-32b5-4051-9ba6-a8dc13f5acf9","Type":"ContainerDied","Data":"29ce9acf7d9b34988b6e12243dae9f32b125b0634a972142f6aa3020b6074fa4"} Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.517873 4846 scope.go:117] "RemoveContainer" containerID="f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.520470 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" event={"ID":"98b51565-aa74-437c-9d42-a93f2b6b52de","Type":"ContainerStarted","Data":"98d0c11360d89106112ce1fba1d9e12a44f52cac83eca4946e2941ab9c3ab55a"} Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.520575 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.566084 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" podStartSLOduration=3.219284323 podStartE2EDuration="11.566061425s" podCreationTimestamp="2026-03-20 11:13:23 +0000 UTC" firstStartedPulling="2026-03-20 11:13:25.031500999 +0000 UTC m=+976.397909235" lastFinishedPulling="2026-03-20 11:13:33.378278101 +0000 UTC m=+984.744686337" observedRunningTime="2026-03-20 11:13:34.563386712 +0000 UTC m=+985.929794968" watchObservedRunningTime="2026-03-20 11:13:34.566061425 +0000 UTC m=+985.932469661" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.580049 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-jjwfh" Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.584645 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:34 crc kubenswrapper[4846]: I0320 11:13:34.589697 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mncwq"] Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.096055 4846 scope.go:117] "RemoveContainer" containerID="dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.124061 4846 scope.go:117] "RemoveContainer" containerID="370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.167135 4846 scope.go:117] "RemoveContainer" containerID="f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b" Mar 20 11:13:35 crc kubenswrapper[4846]: E0320 11:13:35.168335 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b\": container with ID starting with f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b not found: ID does not exist" containerID="f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.168399 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b"} err="failed to get container status \"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b\": rpc error: code = NotFound desc = could not find container \"f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b\": container with ID starting with f0578a66783c9817e6536e091e73643be3dc38aa781a298744ef19193d86667b not found: ID does not exist" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.168429 4846 scope.go:117] "RemoveContainer" containerID="dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65" Mar 20 11:13:35 crc kubenswrapper[4846]: E0320 11:13:35.168885 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65\": container with ID starting with dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65 not found: ID does not exist" containerID="dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.168928 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65"} err="failed to get container status \"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65\": rpc error: code = NotFound desc = could not find container \"dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65\": container with ID starting with dfc37f74196f82c82ae7849eda15e61c649438a8204c316da281c322fb9fce65 not found: ID does not exist" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.168945 4846 scope.go:117] "RemoveContainer" containerID="370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9" Mar 20 11:13:35 crc kubenswrapper[4846]: E0320 11:13:35.169182 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9\": container with ID starting with 370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9 not found: ID does not exist" containerID="370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.169206 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9"} err="failed to get container status \"370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9\": rpc error: code = NotFound desc = could not find container \"370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9\": container with ID starting with 370aa66fa4376a2b2b8f825a03e667a9dc57b330dfe93c728457ff5e8d297bc9 not found: ID does not exist" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.336542 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" path="/var/lib/kubelet/pods/587c0cfb-32b5-4051-9ba6-a8dc13f5acf9/volumes" Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.531100 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c14c61-393f-4234-9d0f-bcc3cf725c6e" containerID="49330526f4320011154892bcc8baeb094b5a73aa777b8a0188fffd000f2f1898" exitCode=0 Mar 20 11:13:35 crc kubenswrapper[4846]: I0320 11:13:35.531206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerDied","Data":"49330526f4320011154892bcc8baeb094b5a73aa777b8a0188fffd000f2f1898"} Mar 20 11:13:36 crc kubenswrapper[4846]: I0320 11:13:36.544062 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c14c61-393f-4234-9d0f-bcc3cf725c6e" containerID="51e8560c4d0726538dc22734872649a9f98fab835ff433a0d7a27623262bebb6" exitCode=0 Mar 20 11:13:36 crc kubenswrapper[4846]: I0320 11:13:36.544224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerDied","Data":"51e8560c4d0726538dc22734872649a9f98fab835ff433a0d7a27623262bebb6"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.369255 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:37 crc kubenswrapper[4846]: E0320 11:13:37.370039 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="registry-server" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.370057 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="registry-server" Mar 20 11:13:37 crc kubenswrapper[4846]: E0320 11:13:37.370082 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="extract-content" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.370090 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="extract-content" Mar 20 11:13:37 crc kubenswrapper[4846]: E0320 11:13:37.370107 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="extract-utilities" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.370114 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="extract-utilities" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.370293 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="587c0cfb-32b5-4051-9ba6-a8dc13f5acf9" containerName="registry-server" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.370849 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.373412 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-cd7n7" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.376419 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.376771 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.397795 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.488816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvk6m\" (UniqueName: \"kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m\") pod \"openstack-operator-index-8p5sx\" (UID: \"1db86603-f12a-4377-8605-bab965d4f407\") " pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.558042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"c351e57dda5e1cd09db3c2ba6984bd065bd771b2f3b7dd7768a4d7dc5c86c4a8"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.559392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"c073733b5af1b83747dfd3aa132b0e68460e40b1362d2797f7655312b1613851"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.559459 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"63b72961e2cdc15996fa91aae01d7e4a7db49e1f8fc63a94c924c1ebdbffd343"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.559551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"8ee1665a8ac0a458cae6bbcf2b0bd497c36c27b24d7a0b58fd3b4002a6f10910"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.559606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"b14c6f489062822ede41e048e6c11026c13f1471081ab6815cb586852a8d443c"} Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.590814 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvk6m\" (UniqueName: \"kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m\") pod \"openstack-operator-index-8p5sx\" (UID: \"1db86603-f12a-4377-8605-bab965d4f407\") " pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.617444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvk6m\" (UniqueName: \"kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m\") pod \"openstack-operator-index-8p5sx\" (UID: \"1db86603-f12a-4377-8605-bab965d4f407\") " pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:37 crc kubenswrapper[4846]: I0320 11:13:37.706393 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.127607 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.570239 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8p5sx" event={"ID":"1db86603-f12a-4377-8605-bab965d4f407","Type":"ContainerStarted","Data":"bf075a1cfea57439f40437ac3488ee2551bbb80dbfc103ec5fbded6238dcd025"} Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.576370 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-d554q" event={"ID":"e6c14c61-393f-4234-9d0f-bcc3cf725c6e","Type":"ContainerStarted","Data":"c7b7903ff4564642789d9555579d89c13f3760af7b4f77a22e8deca975f6584c"} Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.576616 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.614106 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-d554q" podStartSLOduration=6.316577282 podStartE2EDuration="15.614084139s" podCreationTimestamp="2026-03-20 11:13:23 +0000 UTC" firstStartedPulling="2026-03-20 11:13:24.050169406 +0000 UTC m=+975.416577642" lastFinishedPulling="2026-03-20 11:13:33.347676263 +0000 UTC m=+984.714084499" observedRunningTime="2026-03-20 11:13:38.607229876 +0000 UTC m=+989.973638132" watchObservedRunningTime="2026-03-20 11:13:38.614084139 +0000 UTC m=+989.980492375" Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.872255 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:38 crc kubenswrapper[4846]: I0320 11:13:38.920720 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:39 crc kubenswrapper[4846]: I0320 11:13:39.677533 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:13:39 crc kubenswrapper[4846]: I0320 11:13:39.677988 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:13:40 crc kubenswrapper[4846]: I0320 11:13:40.741703 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.346323 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hkt64"] Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.347202 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.359288 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hkt64"] Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.454311 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g4dj\" (UniqueName: \"kubernetes.io/projected/bc323795-0c0f-4463-9b41-f62b2278713a-kube-api-access-9g4dj\") pod \"openstack-operator-index-hkt64\" (UID: \"bc323795-0c0f-4463-9b41-f62b2278713a\") " pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.556113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g4dj\" (UniqueName: \"kubernetes.io/projected/bc323795-0c0f-4463-9b41-f62b2278713a-kube-api-access-9g4dj\") pod \"openstack-operator-index-hkt64\" (UID: \"bc323795-0c0f-4463-9b41-f62b2278713a\") " pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.581683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g4dj\" (UniqueName: \"kubernetes.io/projected/bc323795-0c0f-4463-9b41-f62b2278713a-kube-api-access-9g4dj\") pod \"openstack-operator-index-hkt64\" (UID: \"bc323795-0c0f-4463-9b41-f62b2278713a\") " pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.678261 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:41 crc kubenswrapper[4846]: I0320 11:13:41.910715 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hkt64"] Mar 20 11:13:41 crc kubenswrapper[4846]: W0320 11:13:41.915886 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc323795_0c0f_4463_9b41_f62b2278713a.slice/crio-cb2341adbeba9b703b5efd89b737fde529dda5502e8d20ff37c7ee96c8633ac7 WatchSource:0}: Error finding container cb2341adbeba9b703b5efd89b737fde529dda5502e8d20ff37c7ee96c8633ac7: Status 404 returned error can't find the container with id cb2341adbeba9b703b5efd89b737fde529dda5502e8d20ff37c7ee96c8633ac7 Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.613447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hkt64" event={"ID":"bc323795-0c0f-4463-9b41-f62b2278713a","Type":"ContainerStarted","Data":"489446da01e306fb8488e0abd9efeeca18d853da5afee84d114e373781e665bb"} Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.613884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hkt64" event={"ID":"bc323795-0c0f-4463-9b41-f62b2278713a","Type":"ContainerStarted","Data":"cb2341adbeba9b703b5efd89b737fde529dda5502e8d20ff37c7ee96c8633ac7"} Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.614887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8p5sx" event={"ID":"1db86603-f12a-4377-8605-bab965d4f407","Type":"ContainerStarted","Data":"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0"} Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.615069 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-8p5sx" podUID="1db86603-f12a-4377-8605-bab965d4f407" containerName="registry-server" containerID="cri-o://1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0" gracePeriod=2 Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.637558 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hkt64" podStartSLOduration=1.58593297 podStartE2EDuration="1.637536568s" podCreationTimestamp="2026-03-20 11:13:41 +0000 UTC" firstStartedPulling="2026-03-20 11:13:41.919922187 +0000 UTC m=+993.286330423" lastFinishedPulling="2026-03-20 11:13:41.971525795 +0000 UTC m=+993.337934021" observedRunningTime="2026-03-20 11:13:42.632758515 +0000 UTC m=+993.999166771" watchObservedRunningTime="2026-03-20 11:13:42.637536568 +0000 UTC m=+994.003944804" Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.658640 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8p5sx" podStartSLOduration=2.150229144 podStartE2EDuration="5.65862059s" podCreationTimestamp="2026-03-20 11:13:37 +0000 UTC" firstStartedPulling="2026-03-20 11:13:38.138563928 +0000 UTC m=+989.504972164" lastFinishedPulling="2026-03-20 11:13:41.646955374 +0000 UTC m=+993.013363610" observedRunningTime="2026-03-20 11:13:42.655399824 +0000 UTC m=+994.021808060" watchObservedRunningTime="2026-03-20 11:13:42.65862059 +0000 UTC m=+994.025028826" Mar 20 11:13:42 crc kubenswrapper[4846]: I0320 11:13:42.990858 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.079834 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvk6m\" (UniqueName: \"kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m\") pod \"1db86603-f12a-4377-8605-bab965d4f407\" (UID: \"1db86603-f12a-4377-8605-bab965d4f407\") " Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.086661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m" (OuterVolumeSpecName: "kube-api-access-wvk6m") pod "1db86603-f12a-4377-8605-bab965d4f407" (UID: "1db86603-f12a-4377-8605-bab965d4f407"). InnerVolumeSpecName "kube-api-access-wvk6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.182303 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvk6m\" (UniqueName: \"kubernetes.io/projected/1db86603-f12a-4377-8605-bab965d4f407-kube-api-access-wvk6m\") on node \"crc\" DevicePath \"\"" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.623579 4846 generic.go:334] "Generic (PLEG): container finished" podID="1db86603-f12a-4377-8605-bab965d4f407" containerID="1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0" exitCode=0 Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.624258 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8p5sx" event={"ID":"1db86603-f12a-4377-8605-bab965d4f407","Type":"ContainerDied","Data":"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0"} Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.624303 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8p5sx" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.624340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8p5sx" event={"ID":"1db86603-f12a-4377-8605-bab965d4f407","Type":"ContainerDied","Data":"bf075a1cfea57439f40437ac3488ee2551bbb80dbfc103ec5fbded6238dcd025"} Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.624379 4846 scope.go:117] "RemoveContainer" containerID="1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.650496 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.652764 4846 scope.go:117] "RemoveContainer" containerID="1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.655361 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-8p5sx"] Mar 20 11:13:43 crc kubenswrapper[4846]: E0320 11:13:43.655410 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0\": container with ID starting with 1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0 not found: ID does not exist" containerID="1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0" Mar 20 11:13:43 crc kubenswrapper[4846]: I0320 11:13:43.655487 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0"} err="failed to get container status \"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0\": rpc error: code = NotFound desc = could not find container \"1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0\": container with ID starting with 1042bc07539c3a040170914dab0f3c943c1fd8a2783c16aef84c0523ce263cc0 not found: ID does not exist" Mar 20 11:13:44 crc kubenswrapper[4846]: I0320 11:13:44.463257 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-96pp8" Mar 20 11:13:45 crc kubenswrapper[4846]: I0320 11:13:45.331274 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db86603-f12a-4377-8605-bab965d4f407" path="/var/lib/kubelet/pods/1db86603-f12a-4377-8605-bab965d4f407/volumes" Mar 20 11:13:51 crc kubenswrapper[4846]: I0320 11:13:51.678604 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:51 crc kubenswrapper[4846]: I0320 11:13:51.679037 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:51 crc kubenswrapper[4846]: I0320 11:13:51.708218 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:52 crc kubenswrapper[4846]: I0320 11:13:52.716548 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hkt64" Mar 20 11:13:53 crc kubenswrapper[4846]: I0320 11:13:53.874651 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-d554q" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.975066 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr"] Mar 20 11:13:58 crc kubenswrapper[4846]: E0320 11:13:58.975982 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db86603-f12a-4377-8605-bab965d4f407" containerName="registry-server" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.975996 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db86603-f12a-4377-8605-bab965d4f407" containerName="registry-server" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.976127 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db86603-f12a-4377-8605-bab965d4f407" containerName="registry-server" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.977047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.979016 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5f95x" Mar 20 11:13:58 crc kubenswrapper[4846]: I0320 11:13:58.989409 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr"] Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.021257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.021496 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.021550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc96l\" (UniqueName: \"kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.122550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.122604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc96l\" (UniqueName: \"kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.122648 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.123083 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.123085 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.141751 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc96l\" (UniqueName: \"kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l\") pod \"54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.305693 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.716365 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr"] Mar 20 11:13:59 crc kubenswrapper[4846]: I0320 11:13:59.732310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" event={"ID":"c4028a92-49ba-4337-b2a7-e15686ebf8eb","Type":"ContainerStarted","Data":"5687a3cacd08c6c9f497731b51fe7233095a5b53736411c90c00bc525cba409d"} Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.139491 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566754-fsgs8"] Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.140874 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.143941 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.144977 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.146353 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.149178 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566754-fsgs8"] Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.172543 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcnwm\" (UniqueName: \"kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm\") pod \"auto-csr-approver-29566754-fsgs8\" (UID: \"86e748e6-2de1-464c-ad92-02ad94876039\") " pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.273770 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcnwm\" (UniqueName: \"kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm\") pod \"auto-csr-approver-29566754-fsgs8\" (UID: \"86e748e6-2de1-464c-ad92-02ad94876039\") " pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.293769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcnwm\" (UniqueName: \"kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm\") pod \"auto-csr-approver-29566754-fsgs8\" (UID: \"86e748e6-2de1-464c-ad92-02ad94876039\") " pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.466603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.709124 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566754-fsgs8"] Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.740849 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" event={"ID":"86e748e6-2de1-464c-ad92-02ad94876039","Type":"ContainerStarted","Data":"30b63f4413982417ba177c175ea676211275932895eb761f301b903e75c63387"} Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.742554 4846 generic.go:334] "Generic (PLEG): container finished" podID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerID="30bf30fb258506bafaa68990c3bdbd6dd35fd7b96322c72ca70e63ba1f35263d" exitCode=0 Mar 20 11:14:00 crc kubenswrapper[4846]: I0320 11:14:00.742613 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" event={"ID":"c4028a92-49ba-4337-b2a7-e15686ebf8eb","Type":"ContainerDied","Data":"30bf30fb258506bafaa68990c3bdbd6dd35fd7b96322c72ca70e63ba1f35263d"} Mar 20 11:14:01 crc kubenswrapper[4846]: I0320 11:14:01.751517 4846 generic.go:334] "Generic (PLEG): container finished" podID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerID="071e6151965c2bc7033a94780bd174738477dd2441367363330c89de3d7aef4b" exitCode=0 Mar 20 11:14:01 crc kubenswrapper[4846]: I0320 11:14:01.751602 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" event={"ID":"c4028a92-49ba-4337-b2a7-e15686ebf8eb","Type":"ContainerDied","Data":"071e6151965c2bc7033a94780bd174738477dd2441367363330c89de3d7aef4b"} Mar 20 11:14:02 crc kubenswrapper[4846]: I0320 11:14:02.760103 4846 generic.go:334] "Generic (PLEG): container finished" podID="86e748e6-2de1-464c-ad92-02ad94876039" containerID="71ea7812a6e285a3074f0f4b4ce22be0911fc05882a2d58ee727fd07d8d1beb3" exitCode=0 Mar 20 11:14:02 crc kubenswrapper[4846]: I0320 11:14:02.760155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" event={"ID":"86e748e6-2de1-464c-ad92-02ad94876039","Type":"ContainerDied","Data":"71ea7812a6e285a3074f0f4b4ce22be0911fc05882a2d58ee727fd07d8d1beb3"} Mar 20 11:14:02 crc kubenswrapper[4846]: I0320 11:14:02.763574 4846 generic.go:334] "Generic (PLEG): container finished" podID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerID="edf75a80bdc4820f1ae750a111ac85dd525493de8ce3ad2afa2bd1e65be69c34" exitCode=0 Mar 20 11:14:02 crc kubenswrapper[4846]: I0320 11:14:02.763616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" event={"ID":"c4028a92-49ba-4337-b2a7-e15686ebf8eb","Type":"ContainerDied","Data":"edf75a80bdc4820f1ae750a111ac85dd525493de8ce3ad2afa2bd1e65be69c34"} Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.120874 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.132980 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.233493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcnwm\" (UniqueName: \"kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm\") pod \"86e748e6-2de1-464c-ad92-02ad94876039\" (UID: \"86e748e6-2de1-464c-ad92-02ad94876039\") " Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.239848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm" (OuterVolumeSpecName: "kube-api-access-kcnwm") pod "86e748e6-2de1-464c-ad92-02ad94876039" (UID: "86e748e6-2de1-464c-ad92-02ad94876039"). InnerVolumeSpecName "kube-api-access-kcnwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.335837 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle\") pod \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.336017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util\") pod \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.336093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc96l\" (UniqueName: \"kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l\") pod \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\" (UID: \"c4028a92-49ba-4337-b2a7-e15686ebf8eb\") " Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.336719 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle" (OuterVolumeSpecName: "bundle") pod "c4028a92-49ba-4337-b2a7-e15686ebf8eb" (UID: "c4028a92-49ba-4337-b2a7-e15686ebf8eb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.337072 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.337092 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcnwm\" (UniqueName: \"kubernetes.io/projected/86e748e6-2de1-464c-ad92-02ad94876039-kube-api-access-kcnwm\") on node \"crc\" DevicePath \"\"" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.339298 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l" (OuterVolumeSpecName: "kube-api-access-sc96l") pod "c4028a92-49ba-4337-b2a7-e15686ebf8eb" (UID: "c4028a92-49ba-4337-b2a7-e15686ebf8eb"). InnerVolumeSpecName "kube-api-access-sc96l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.349044 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util" (OuterVolumeSpecName: "util") pod "c4028a92-49ba-4337-b2a7-e15686ebf8eb" (UID: "c4028a92-49ba-4337-b2a7-e15686ebf8eb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.437877 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c4028a92-49ba-4337-b2a7-e15686ebf8eb-util\") on node \"crc\" DevicePath \"\"" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.437927 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc96l\" (UniqueName: \"kubernetes.io/projected/c4028a92-49ba-4337-b2a7-e15686ebf8eb-kube-api-access-sc96l\") on node \"crc\" DevicePath \"\"" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.777078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.777694 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566754-fsgs8" event={"ID":"86e748e6-2de1-464c-ad92-02ad94876039","Type":"ContainerDied","Data":"30b63f4413982417ba177c175ea676211275932895eb761f301b903e75c63387"} Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.777776 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b63f4413982417ba177c175ea676211275932895eb761f301b903e75c63387" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.780207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" event={"ID":"c4028a92-49ba-4337-b2a7-e15686ebf8eb","Type":"ContainerDied","Data":"5687a3cacd08c6c9f497731b51fe7233095a5b53736411c90c00bc525cba409d"} Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.780234 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5687a3cacd08c6c9f497731b51fe7233095a5b53736411c90c00bc525cba409d" Mar 20 11:14:04 crc kubenswrapper[4846]: I0320 11:14:04.780277 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr" Mar 20 11:14:05 crc kubenswrapper[4846]: I0320 11:14:05.191587 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566748-9t2fn"] Mar 20 11:14:05 crc kubenswrapper[4846]: I0320 11:14:05.197553 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566748-9t2fn"] Mar 20 11:14:05 crc kubenswrapper[4846]: I0320 11:14:05.332315 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="456687eb-df67-4f43-81ef-4c7d146cc813" path="/var/lib/kubelet/pods/456687eb-df67-4f43-81ef-4c7d146cc813/volumes" Mar 20 11:14:09 crc kubenswrapper[4846]: I0320 11:14:09.677654 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:14:09 crc kubenswrapper[4846]: I0320 11:14:09.678273 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695092 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94"] Mar 20 11:14:12 crc kubenswrapper[4846]: E0320 11:14:12.695426 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="pull" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695443 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="pull" Mar 20 11:14:12 crc kubenswrapper[4846]: E0320 11:14:12.695459 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86e748e6-2de1-464c-ad92-02ad94876039" containerName="oc" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695466 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="86e748e6-2de1-464c-ad92-02ad94876039" containerName="oc" Mar 20 11:14:12 crc kubenswrapper[4846]: E0320 11:14:12.695476 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="util" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695483 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="util" Mar 20 11:14:12 crc kubenswrapper[4846]: E0320 11:14:12.695496 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="extract" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695502 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="extract" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695614 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="86e748e6-2de1-464c-ad92-02ad94876039" containerName="oc" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.695626 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4028a92-49ba-4337-b2a7-e15686ebf8eb" containerName="extract" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.696185 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.704041 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-g2h9d" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.719992 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94"] Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.752203 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsxzt\" (UniqueName: \"kubernetes.io/projected/4b359265-9a42-4e56-a6ea-108fafb33ae3-kube-api-access-vsxzt\") pod \"openstack-operator-controller-init-7fc6f9b5fc-djp94\" (UID: \"4b359265-9a42-4e56-a6ea-108fafb33ae3\") " pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.854062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsxzt\" (UniqueName: \"kubernetes.io/projected/4b359265-9a42-4e56-a6ea-108fafb33ae3-kube-api-access-vsxzt\") pod \"openstack-operator-controller-init-7fc6f9b5fc-djp94\" (UID: \"4b359265-9a42-4e56-a6ea-108fafb33ae3\") " pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:12 crc kubenswrapper[4846]: I0320 11:14:12.878597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsxzt\" (UniqueName: \"kubernetes.io/projected/4b359265-9a42-4e56-a6ea-108fafb33ae3-kube-api-access-vsxzt\") pod \"openstack-operator-controller-init-7fc6f9b5fc-djp94\" (UID: \"4b359265-9a42-4e56-a6ea-108fafb33ae3\") " pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:13 crc kubenswrapper[4846]: I0320 11:14:13.023101 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:13 crc kubenswrapper[4846]: I0320 11:14:13.292398 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94"] Mar 20 11:14:13 crc kubenswrapper[4846]: I0320 11:14:13.830155 4846 scope.go:117] "RemoveContainer" containerID="243e667f4ac6502ea55aaa409785e02a174f6d47feaa7a7767a7546bcdd4f6d8" Mar 20 11:14:13 crc kubenswrapper[4846]: I0320 11:14:13.856733 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" event={"ID":"4b359265-9a42-4e56-a6ea-108fafb33ae3","Type":"ContainerStarted","Data":"ee0f82cb77915ff4bedf8cbfccf04b152e4f0cb13eb50e34dd98632f29536e93"} Mar 20 11:14:18 crc kubenswrapper[4846]: I0320 11:14:18.892018 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" event={"ID":"4b359265-9a42-4e56-a6ea-108fafb33ae3","Type":"ContainerStarted","Data":"ebcc5d827e7c7f62abbbd75f995107d04a179618cac904287a86cb18d99c8f17"} Mar 20 11:14:18 crc kubenswrapper[4846]: I0320 11:14:18.892738 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:18 crc kubenswrapper[4846]: I0320 11:14:18.923305 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" podStartSLOduration=1.8793818199999999 podStartE2EDuration="6.923286604s" podCreationTimestamp="2026-03-20 11:14:12 +0000 UTC" firstStartedPulling="2026-03-20 11:14:13.298515702 +0000 UTC m=+1024.664923938" lastFinishedPulling="2026-03-20 11:14:18.342420486 +0000 UTC m=+1029.708828722" observedRunningTime="2026-03-20 11:14:18.920691442 +0000 UTC m=+1030.287099698" watchObservedRunningTime="2026-03-20 11:14:18.923286604 +0000 UTC m=+1030.289694830" Mar 20 11:14:23 crc kubenswrapper[4846]: I0320 11:14:23.027271 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-7fc6f9b5fc-djp94" Mar 20 11:14:39 crc kubenswrapper[4846]: I0320 11:14:39.678175 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:14:39 crc kubenswrapper[4846]: I0320 11:14:39.678974 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:14:39 crc kubenswrapper[4846]: I0320 11:14:39.679039 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:14:39 crc kubenswrapper[4846]: I0320 11:14:39.679825 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:14:39 crc kubenswrapper[4846]: I0320 11:14:39.679917 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45" gracePeriod=600 Mar 20 11:14:40 crc kubenswrapper[4846]: I0320 11:14:40.038422 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45" exitCode=0 Mar 20 11:14:40 crc kubenswrapper[4846]: I0320 11:14:40.038487 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45"} Mar 20 11:14:40 crc kubenswrapper[4846]: I0320 11:14:40.038881 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2"} Mar 20 11:14:40 crc kubenswrapper[4846]: I0320 11:14:40.038925 4846 scope.go:117] "RemoveContainer" containerID="0a57e56ab9d8df8651ed3ea975273e74925c5f6b2165f31a199a7682002a6d57" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.344241 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.347356 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.381679 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.482332 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.482438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.482473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkcpd\" (UniqueName: \"kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.584177 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.584579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkcpd\" (UniqueName: \"kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.584738 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.584939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.585459 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.621676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkcpd\" (UniqueName: \"kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd\") pod \"certified-operators-4622r\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:53 crc kubenswrapper[4846]: I0320 11:14:53.671716 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:14:54 crc kubenswrapper[4846]: I0320 11:14:54.094534 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:14:54 crc kubenswrapper[4846]: I0320 11:14:54.157994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerStarted","Data":"6f9eb2b36e70fb6093481b027d62b010f716d9bd9d170d323cdefa93aee264b5"} Mar 20 11:14:55 crc kubenswrapper[4846]: I0320 11:14:55.169166 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerID="9086ed4556a5802cbd4d6f2f83c978521bc4e3ae1b43755ec53a2ada4b7cdb15" exitCode=0 Mar 20 11:14:55 crc kubenswrapper[4846]: I0320 11:14:55.169230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerDied","Data":"9086ed4556a5802cbd4d6f2f83c978521bc4e3ae1b43755ec53a2ada4b7cdb15"} Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.194946 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerStarted","Data":"8e4b497960aa01e3270d628196087fbfd0a7eb853b9fe0f95f6a0ae8c6ce790e"} Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.353238 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.354660 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.360157 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fdfx6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.361504 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.362422 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.367533 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-l4ww6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.370506 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.374733 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.411209 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-ht246"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.412594 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.414686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-ckzlj" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.415091 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.416027 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.422228 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-48dtb" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.422492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-ht246"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.436635 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.447975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v78ks\" (UniqueName: \"kubernetes.io/projected/aff6e349-68ea-4f3c-8514-b2a7c7ffed23-kube-api-access-v78ks\") pod \"cinder-operator-controller-manager-8d58dc466-ggrpk\" (UID: \"aff6e349-68ea-4f3c-8514-b2a7c7ffed23\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.448139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npsp8\" (UniqueName: \"kubernetes.io/projected/529816a9-1274-4a4f-88b2-6e0b426075a0-kube-api-access-npsp8\") pod \"barbican-operator-controller-manager-59bc569d95-zncs8\" (UID: \"529816a9-1274-4a4f-88b2-6e0b426075a0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.489293 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.490386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.504582 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-pfjvx" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.523940 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.535270 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.536363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.544938 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zqrqd" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.550594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npsp8\" (UniqueName: \"kubernetes.io/projected/529816a9-1274-4a4f-88b2-6e0b426075a0-kube-api-access-npsp8\") pod \"barbican-operator-controller-manager-59bc569d95-zncs8\" (UID: \"529816a9-1274-4a4f-88b2-6e0b426075a0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.550659 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lpvc\" (UniqueName: \"kubernetes.io/projected/24727b45-3896-4b9f-a91e-62434aa32ac7-kube-api-access-7lpvc\") pod \"heat-operator-controller-manager-67dd5f86f5-thhs6\" (UID: \"24727b45-3896-4b9f-a91e-62434aa32ac7\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.550694 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v78ks\" (UniqueName: \"kubernetes.io/projected/aff6e349-68ea-4f3c-8514-b2a7c7ffed23-kube-api-access-v78ks\") pod \"cinder-operator-controller-manager-8d58dc466-ggrpk\" (UID: \"aff6e349-68ea-4f3c-8514-b2a7c7ffed23\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.550736 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jc87\" (UniqueName: \"kubernetes.io/projected/380b5a6a-c151-4763-a86b-c8e53b6f1aa2-kube-api-access-5jc87\") pod \"glance-operator-controller-manager-79df6bcc97-4m4lg\" (UID: \"380b5a6a-c151-4763-a86b-c8e53b6f1aa2\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.550764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hdgj\" (UniqueName: \"kubernetes.io/projected/9bf60e8a-5221-44ba-8471-5b9e2692d2a7-kube-api-access-2hdgj\") pod \"designate-operator-controller-manager-588d4d986b-ht246\" (UID: \"9bf60e8a-5221-44ba-8471-5b9e2692d2a7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.562443 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.592492 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npsp8\" (UniqueName: \"kubernetes.io/projected/529816a9-1274-4a4f-88b2-6e0b426075a0-kube-api-access-npsp8\") pod \"barbican-operator-controller-manager-59bc569d95-zncs8\" (UID: \"529816a9-1274-4a4f-88b2-6e0b426075a0\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.594269 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v78ks\" (UniqueName: \"kubernetes.io/projected/aff6e349-68ea-4f3c-8514-b2a7c7ffed23-kube-api-access-v78ks\") pod \"cinder-operator-controller-manager-8d58dc466-ggrpk\" (UID: \"aff6e349-68ea-4f3c-8514-b2a7c7ffed23\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.599361 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.612369 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.617136 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r5shd" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.617361 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.633969 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.643201 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.644609 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.647094 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-brq8z" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654478 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8gzf\" (UniqueName: \"kubernetes.io/projected/6be091f8-9109-4fce-a856-feacf04c76d4-kube-api-access-r8gzf\") pod \"horizon-operator-controller-manager-8464cc45fb-qg5jf\" (UID: \"6be091f8-9109-4fce-a856-feacf04c76d4\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654628 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lpvc\" (UniqueName: \"kubernetes.io/projected/24727b45-3896-4b9f-a91e-62434aa32ac7-kube-api-access-7lpvc\") pod \"heat-operator-controller-manager-67dd5f86f5-thhs6\" (UID: \"24727b45-3896-4b9f-a91e-62434aa32ac7\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jc87\" (UniqueName: \"kubernetes.io/projected/380b5a6a-c151-4763-a86b-c8e53b6f1aa2-kube-api-access-5jc87\") pod \"glance-operator-controller-manager-79df6bcc97-4m4lg\" (UID: \"380b5a6a-c151-4763-a86b-c8e53b6f1aa2\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654690 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7pt2\" (UniqueName: \"kubernetes.io/projected/1132b3dc-feeb-41ce-a579-d5b089b86597-kube-api-access-w7pt2\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.654714 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hdgj\" (UniqueName: \"kubernetes.io/projected/9bf60e8a-5221-44ba-8471-5b9e2692d2a7-kube-api-access-2hdgj\") pod \"designate-operator-controller-manager-588d4d986b-ht246\" (UID: \"9bf60e8a-5221-44ba-8471-5b9e2692d2a7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.693104 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.695793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jc87\" (UniqueName: \"kubernetes.io/projected/380b5a6a-c151-4763-a86b-c8e53b6f1aa2-kube-api-access-5jc87\") pod \"glance-operator-controller-manager-79df6bcc97-4m4lg\" (UID: \"380b5a6a-c151-4763-a86b-c8e53b6f1aa2\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.707781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hdgj\" (UniqueName: \"kubernetes.io/projected/9bf60e8a-5221-44ba-8471-5b9e2692d2a7-kube-api-access-2hdgj\") pod \"designate-operator-controller-manager-588d4d986b-ht246\" (UID: \"9bf60e8a-5221-44ba-8471-5b9e2692d2a7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.711084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lpvc\" (UniqueName: \"kubernetes.io/projected/24727b45-3896-4b9f-a91e-62434aa32ac7-kube-api-access-7lpvc\") pod \"heat-operator-controller-manager-67dd5f86f5-thhs6\" (UID: \"24727b45-3896-4b9f-a91e-62434aa32ac7\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.724336 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.729006 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.729261 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.743333 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-cnbsg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.756481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8gzf\" (UniqueName: \"kubernetes.io/projected/6be091f8-9109-4fce-a856-feacf04c76d4-kube-api-access-r8gzf\") pod \"horizon-operator-controller-manager-8464cc45fb-qg5jf\" (UID: \"6be091f8-9109-4fce-a856-feacf04c76d4\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.756528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.756577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsf77\" (UniqueName: \"kubernetes.io/projected/c88c55bf-7869-40e5-90f7-0bfdcb676b9e-kube-api-access-lsf77\") pod \"ironic-operator-controller-manager-6f787dddc9-g8hzn\" (UID: \"c88c55bf-7869-40e5-90f7-0bfdcb676b9e\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.756652 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7pt2\" (UniqueName: \"kubernetes.io/projected/1132b3dc-feeb-41ce-a579-d5b089b86597-kube-api-access-w7pt2\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: E0320 11:14:57.757168 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:57 crc kubenswrapper[4846]: E0320 11:14:57.757218 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:14:58.257198455 +0000 UTC m=+1069.623606691 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.757494 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.765586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.780444 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.789462 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.789706 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.790885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.797618 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2srjl" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.800423 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.802969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.826374 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9ghdl" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.834677 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7pt2\" (UniqueName: \"kubernetes.io/projected/1132b3dc-feeb-41ce-a579-d5b089b86597-kube-api-access-w7pt2\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.838181 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.849031 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.865725 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8gzf\" (UniqueName: \"kubernetes.io/projected/6be091f8-9109-4fce-a856-feacf04c76d4-kube-api-access-r8gzf\") pod \"horizon-operator-controller-manager-8464cc45fb-qg5jf\" (UID: \"6be091f8-9109-4fce-a856-feacf04c76d4\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.866643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.867310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsf77\" (UniqueName: \"kubernetes.io/projected/c88c55bf-7869-40e5-90f7-0bfdcb676b9e-kube-api-access-lsf77\") pod \"ironic-operator-controller-manager-6f787dddc9-g8hzn\" (UID: \"c88c55bf-7869-40e5-90f7-0bfdcb676b9e\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.867377 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lql5b\" (UniqueName: \"kubernetes.io/projected/0b70e378-a53e-4519-b415-64810a544874-kube-api-access-lql5b\") pod \"mariadb-operator-controller-manager-67ccfc9778-r7pgd\" (UID: \"0b70e378-a53e-4519-b415-64810a544874\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.867412 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/2bba6588-37b3-453e-972e-393208ddd560-kube-api-access-qn9ck\") pod \"keystone-operator-controller-manager-768b96df4c-qwptc\" (UID: \"2bba6588-37b3-453e-972e-393208ddd560\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.867444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkb6k\" (UniqueName: \"kubernetes.io/projected/86d13cbe-cc30-41b6-8829-1f5cbb99c48c-kube-api-access-kkb6k\") pod \"manila-operator-controller-manager-55f864c847-hgqt9\" (UID: \"86d13cbe-cc30-41b6-8829-1f5cbb99c48c\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.895468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsf77\" (UniqueName: \"kubernetes.io/projected/c88c55bf-7869-40e5-90f7-0bfdcb676b9e-kube-api-access-lsf77\") pod \"ironic-operator-controller-manager-6f787dddc9-g8hzn\" (UID: \"c88c55bf-7869-40e5-90f7-0bfdcb676b9e\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.898885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.909186 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.969038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkb6k\" (UniqueName: \"kubernetes.io/projected/86d13cbe-cc30-41b6-8829-1f5cbb99c48c-kube-api-access-kkb6k\") pod \"manila-operator-controller-manager-55f864c847-hgqt9\" (UID: \"86d13cbe-cc30-41b6-8829-1f5cbb99c48c\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.969186 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lql5b\" (UniqueName: \"kubernetes.io/projected/0b70e378-a53e-4519-b415-64810a544874-kube-api-access-lql5b\") pod \"mariadb-operator-controller-manager-67ccfc9778-r7pgd\" (UID: \"0b70e378-a53e-4519-b415-64810a544874\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.969228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/2bba6588-37b3-453e-972e-393208ddd560-kube-api-access-qn9ck\") pod \"keystone-operator-controller-manager-768b96df4c-qwptc\" (UID: \"2bba6588-37b3-453e-972e-393208ddd560\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.974538 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.975575 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6"] Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.975714 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.976272 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.979275 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.979746 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ppmqs" Mar 20 11:14:57 crc kubenswrapper[4846]: I0320 11:14:57.980706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-blcmn" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.001112 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.036840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.038014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkb6k\" (UniqueName: \"kubernetes.io/projected/86d13cbe-cc30-41b6-8829-1f5cbb99c48c-kube-api-access-kkb6k\") pod \"manila-operator-controller-manager-55f864c847-hgqt9\" (UID: \"86d13cbe-cc30-41b6-8829-1f5cbb99c48c\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.059481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lql5b\" (UniqueName: \"kubernetes.io/projected/0b70e378-a53e-4519-b415-64810a544874-kube-api-access-lql5b\") pod \"mariadb-operator-controller-manager-67ccfc9778-r7pgd\" (UID: \"0b70e378-a53e-4519-b415-64810a544874\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.064359 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.066564 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-d8fqt" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.071444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkfsx\" (UniqueName: \"kubernetes.io/projected/06eb056c-f3f0-4242-bbee-d201dfd37925-kube-api-access-dkfsx\") pod \"octavia-operator-controller-manager-5b9f45d989-rrx9r\" (UID: \"06eb056c-f3f0-4242-bbee-d201dfd37925\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.071502 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxhv7\" (UniqueName: \"kubernetes.io/projected/bd21bfeb-d2b2-4757-aa36-094c039951f0-kube-api-access-sxhv7\") pod \"nova-operator-controller-manager-5d488d59fb-f45j6\" (UID: \"bd21bfeb-d2b2-4757-aa36-094c039951f0\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.071647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxj26\" (UniqueName: \"kubernetes.io/projected/fa16f94f-be9c-4e2c-8dc3-c420db3ed639-kube-api-access-qxj26\") pod \"neutron-operator-controller-manager-767865f676-9wzfn\" (UID: \"fa16f94f-be9c-4e2c-8dc3-c420db3ed639\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.073253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn9ck\" (UniqueName: \"kubernetes.io/projected/2bba6588-37b3-453e-972e-393208ddd560-kube-api-access-qn9ck\") pod \"keystone-operator-controller-manager-768b96df4c-qwptc\" (UID: \"2bba6588-37b3-453e-972e-393208ddd560\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.096290 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.106582 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.111229 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.120830 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.122167 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.122295 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.122970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.128752 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.140010 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.129443 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-bmjzc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.129510 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cvb79" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.170095 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxhv7\" (UniqueName: \"kubernetes.io/projected/bd21bfeb-d2b2-4757-aa36-094c039951f0-kube-api-access-sxhv7\") pod \"nova-operator-controller-manager-5d488d59fb-f45j6\" (UID: \"bd21bfeb-d2b2-4757-aa36-094c039951f0\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172205 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb554\" (UniqueName: \"kubernetes.io/projected/7d330820-2aac-4076-a205-8dedd331cae1-kube-api-access-kb554\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxj26\" (UniqueName: \"kubernetes.io/projected/fa16f94f-be9c-4e2c-8dc3-c420db3ed639-kube-api-access-qxj26\") pod \"neutron-operator-controller-manager-767865f676-9wzfn\" (UID: \"fa16f94f-be9c-4e2c-8dc3-c420db3ed639\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4tbv\" (UniqueName: \"kubernetes.io/projected/64ce367c-20db-4598-8006-f51d8aa22bdd-kube-api-access-n4tbv\") pod \"ovn-operator-controller-manager-884679f54-gqbrp\" (UID: \"64ce367c-20db-4598-8006-f51d8aa22bdd\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.172377 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkfsx\" (UniqueName: \"kubernetes.io/projected/06eb056c-f3f0-4242-bbee-d201dfd37925-kube-api-access-dkfsx\") pod \"octavia-operator-controller-manager-5b9f45d989-rrx9r\" (UID: \"06eb056c-f3f0-4242-bbee-d201dfd37925\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.179595 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-4fww7"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.180700 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.185413 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.194451 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.204608 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.207923 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.210960 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lrwx6" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.211301 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jxfw7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.235156 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-4fww7"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.247712 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxj26\" (UniqueName: \"kubernetes.io/projected/fa16f94f-be9c-4e2c-8dc3-c420db3ed639-kube-api-access-qxj26\") pod \"neutron-operator-controller-manager-767865f676-9wzfn\" (UID: \"fa16f94f-be9c-4e2c-8dc3-c420db3ed639\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.258204 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkfsx\" (UniqueName: \"kubernetes.io/projected/06eb056c-f3f0-4242-bbee-d201dfd37925-kube-api-access-dkfsx\") pod \"octavia-operator-controller-manager-5b9f45d989-rrx9r\" (UID: \"06eb056c-f3f0-4242-bbee-d201dfd37925\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.266756 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxhv7\" (UniqueName: \"kubernetes.io/projected/bd21bfeb-d2b2-4757-aa36-094c039951f0-kube-api-access-sxhv7\") pod \"nova-operator-controller-manager-5d488d59fb-f45j6\" (UID: \"bd21bfeb-d2b2-4757-aa36-094c039951f0\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.290441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4tbv\" (UniqueName: \"kubernetes.io/projected/64ce367c-20db-4598-8006-f51d8aa22bdd-kube-api-access-n4tbv\") pod \"ovn-operator-controller-manager-884679f54-gqbrp\" (UID: \"64ce367c-20db-4598-8006-f51d8aa22bdd\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.290602 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.290653 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.290730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb554\" (UniqueName: \"kubernetes.io/projected/7d330820-2aac-4076-a205-8dedd331cae1-kube-api-access-kb554\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.292614 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.297084 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:14:58.797052545 +0000 UTC m=+1070.163460781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.292681 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.297268 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:14:59.297256679 +0000 UTC m=+1070.663664915 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.324354 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerID="8e4b497960aa01e3270d628196087fbfd0a7eb853b9fe0f95f6a0ae8c6ce790e" exitCode=0 Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.324435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerDied","Data":"8e4b497960aa01e3270d628196087fbfd0a7eb853b9fe0f95f6a0ae8c6ce790e"} Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.349997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb554\" (UniqueName: \"kubernetes.io/projected/7d330820-2aac-4076-a205-8dedd331cae1-kube-api-access-kb554\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.351864 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4tbv\" (UniqueName: \"kubernetes.io/projected/64ce367c-20db-4598-8006-f51d8aa22bdd-kube-api-access-n4tbv\") pod \"ovn-operator-controller-manager-884679f54-gqbrp\" (UID: \"64ce367c-20db-4598-8006-f51d8aa22bdd\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.388214 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.393975 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.406958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.409102 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.410214 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh2hx\" (UniqueName: \"kubernetes.io/projected/132090d4-b1d3-485e-9525-26fe1eecc448-kube-api-access-dh2hx\") pod \"placement-operator-controller-manager-5784578c99-dr5ms\" (UID: \"132090d4-b1d3-485e-9525-26fe1eecc448\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.410257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhnpw\" (UniqueName: \"kubernetes.io/projected/9c7c760d-4ef7-41a0-bca6-c9b26d99fd52-kube-api-access-jhnpw\") pod \"swift-operator-controller-manager-c674c5965-4fww7\" (UID: \"9c7c760d-4ef7-41a0-bca6-c9b26d99fd52\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.427336 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-l9kzq" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.473933 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.483995 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.519117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhnpw\" (UniqueName: \"kubernetes.io/projected/9c7c760d-4ef7-41a0-bca6-c9b26d99fd52-kube-api-access-jhnpw\") pod \"swift-operator-controller-manager-c674c5965-4fww7\" (UID: \"9c7c760d-4ef7-41a0-bca6-c9b26d99fd52\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.519248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lthg\" (UniqueName: \"kubernetes.io/projected/14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd-kube-api-access-7lthg\") pod \"telemetry-operator-controller-manager-d6b694c5-ppjfc\" (UID: \"14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.519635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh2hx\" (UniqueName: \"kubernetes.io/projected/132090d4-b1d3-485e-9525-26fe1eecc448-kube-api-access-dh2hx\") pod \"placement-operator-controller-manager-5784578c99-dr5ms\" (UID: \"132090d4-b1d3-485e-9525-26fe1eecc448\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.530665 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.548231 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.557001 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.565515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhnpw\" (UniqueName: \"kubernetes.io/projected/9c7c760d-4ef7-41a0-bca6-c9b26d99fd52-kube-api-access-jhnpw\") pod \"swift-operator-controller-manager-c674c5965-4fww7\" (UID: \"9c7c760d-4ef7-41a0-bca6-c9b26d99fd52\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.566075 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qgh24" Mar 20 11:14:58 crc kubenswrapper[4846]: W0320 11:14:58.569747 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod529816a9_1274_4a4f_88b2_6e0b426075a0.slice/crio-2c0daaeb6ebeef680e1b416913bfed045d133121313d5539e0dbd715ac8e18da WatchSource:0}: Error finding container 2c0daaeb6ebeef680e1b416913bfed045d133121313d5539e0dbd715ac8e18da: Status 404 returned error can't find the container with id 2c0daaeb6ebeef680e1b416913bfed045d133121313d5539e0dbd715ac8e18da Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.595948 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.597233 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.598215 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.603512 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-78zgg" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.604932 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.615093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh2hx\" (UniqueName: \"kubernetes.io/projected/132090d4-b1d3-485e-9525-26fe1eecc448-kube-api-access-dh2hx\") pod \"placement-operator-controller-manager-5784578c99-dr5ms\" (UID: \"132090d4-b1d3-485e-9525-26fe1eecc448\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.622108 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lthg\" (UniqueName: \"kubernetes.io/projected/14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd-kube-api-access-7lthg\") pod \"telemetry-operator-controller-manager-d6b694c5-ppjfc\" (UID: \"14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.622215 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf72p\" (UniqueName: \"kubernetes.io/projected/e7f86e9b-89bb-4558-aa0d-4bec37131052-kube-api-access-mf72p\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8ng9\" (UID: \"e7f86e9b-89bb-4558-aa0d-4bec37131052\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.622257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjbhm\" (UniqueName: \"kubernetes.io/projected/b155d1b3-b553-414b-9017-329ee45a9658-kube-api-access-vjbhm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-dvj5j\" (UID: \"b155d1b3-b553-414b-9017-329ee45a9658\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.655431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.675223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lthg\" (UniqueName: \"kubernetes.io/projected/14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd-kube-api-access-7lthg\") pod \"telemetry-operator-controller-manager-d6b694c5-ppjfc\" (UID: \"14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.686213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.704174 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.705271 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.706512 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.710908 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.710927 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.721930 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sxrj5" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.722872 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjbhm\" (UniqueName: \"kubernetes.io/projected/b155d1b3-b553-414b-9017-329ee45a9658-kube-api-access-vjbhm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-dvj5j\" (UID: \"b155d1b3-b553-414b-9017-329ee45a9658\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.722969 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhhbg\" (UniqueName: \"kubernetes.io/projected/de215fcd-2be6-40ca-8ae1-d628b6b4280b-kube-api-access-bhhbg\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.723063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.723090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf72p\" (UniqueName: \"kubernetes.io/projected/e7f86e9b-89bb-4558-aa0d-4bec37131052-kube-api-access-mf72p\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8ng9\" (UID: \"e7f86e9b-89bb-4558-aa0d-4bec37131052\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.723111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.758219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjbhm\" (UniqueName: \"kubernetes.io/projected/b155d1b3-b553-414b-9017-329ee45a9658-kube-api-access-vjbhm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-dvj5j\" (UID: \"b155d1b3-b553-414b-9017-329ee45a9658\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.759205 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf72p\" (UniqueName: \"kubernetes.io/projected/e7f86e9b-89bb-4558-aa0d-4bec37131052-kube-api-access-mf72p\") pod \"test-operator-controller-manager-5c5cb9c4d7-b8ng9\" (UID: \"e7f86e9b-89bb-4558-aa0d-4bec37131052\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.822910 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.823732 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.824701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.824798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.824837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.824913 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhhbg\" (UniqueName: \"kubernetes.io/projected/de215fcd-2be6-40ca-8ae1-d628b6b4280b-kube-api-access-bhhbg\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825252 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825368 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:14:59.325343117 +0000 UTC m=+1070.691751353 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825402 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825512 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:14:59.325504221 +0000 UTC m=+1070.691912457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825457 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: E0320 11:14:58.825651 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:14:59.825640115 +0000 UTC m=+1071.192048351 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.841592 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.844614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhhbg\" (UniqueName: \"kubernetes.io/projected/de215fcd-2be6-40ca-8ae1-d628b6b4280b-kube-api-access-bhhbg\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.862799 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg"] Mar 20 11:14:58 crc kubenswrapper[4846]: I0320 11:14:58.982277 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.001302 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.065833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.341850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.342334 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.342415 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:01.342392482 +0000 UTC m=+1072.708800718 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.342443 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.342497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.342618 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.342883 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.343066 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:00.343014297 +0000 UTC m=+1071.709422533 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.345286 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:00.345267831 +0000 UTC m=+1071.711676067 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.378986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" event={"ID":"529816a9-1274-4a4f-88b2-6e0b426075a0","Type":"ContainerStarted","Data":"2c0daaeb6ebeef680e1b416913bfed045d133121313d5539e0dbd715ac8e18da"} Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.387315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" event={"ID":"24727b45-3896-4b9f-a91e-62434aa32ac7","Type":"ContainerStarted","Data":"89daea040f22509026605e87151d0c648807741dfc870885b98af4241a9988e2"} Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.408206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" event={"ID":"aff6e349-68ea-4f3c-8514-b2a7c7ffed23","Type":"ContainerStarted","Data":"5bacf05d08073e36b1c7b2db7d8313ed6da759e98930e98b3f0079da3b2967cd"} Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.420651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" event={"ID":"380b5a6a-c151-4763-a86b-c8e53b6f1aa2","Type":"ContainerStarted","Data":"a02e5b59173c6d6e35dab51a1fff9d76ba7db3b4c1fc63c96014f9af34570786"} Mar 20 11:14:59 crc kubenswrapper[4846]: W0320 11:14:59.512222 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bf60e8a_5221_44ba_8471_5b9e2692d2a7.slice/crio-62f16d9c06756da4eb9f30177f7caf671ab5f66668ce620d5a528b4602c3b231 WatchSource:0}: Error finding container 62f16d9c06756da4eb9f30177f7caf671ab5f66668ce620d5a528b4602c3b231: Status 404 returned error can't find the container with id 62f16d9c06756da4eb9f30177f7caf671ab5f66668ce620d5a528b4602c3b231 Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.512664 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-ht246"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.529109 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.682219 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.741835 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.790534 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.797100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.854536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.855157 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: E0320 11:14:59.855242 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:01.855216645 +0000 UTC m=+1073.221624881 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.923780 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.932958 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-4fww7"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.939416 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc"] Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.947387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn"] Mar 20 11:14:59 crc kubenswrapper[4846]: W0320 11:14:59.952118 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06eb056c_f3f0_4242_bbee_d201dfd37925.slice/crio-5f434b330658fb21de2108a46beae4a5030cf82a1e12b0151d70555edd1cad5d WatchSource:0}: Error finding container 5f434b330658fb21de2108a46beae4a5030cf82a1e12b0151d70555edd1cad5d: Status 404 returned error can't find the container with id 5f434b330658fb21de2108a46beae4a5030cf82a1e12b0151d70555edd1cad5d Mar 20 11:14:59 crc kubenswrapper[4846]: I0320 11:14:59.961411 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6"] Mar 20 11:14:59 crc kubenswrapper[4846]: W0320 11:14:59.979121 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14ce4a99_f370_4aaa_bb8f_bd3f1d8746dd.slice/crio-a58a42628874036dd493e29c8305b4f6dab95370f9fba4472cf7f50082aeea0f WatchSource:0}: Error finding container a58a42628874036dd493e29c8305b4f6dab95370f9fba4472cf7f50082aeea0f: Status 404 returned error can't find the container with id a58a42628874036dd493e29c8305b4f6dab95370f9fba4472cf7f50082aeea0f Mar 20 11:14:59 crc kubenswrapper[4846]: W0320 11:14:59.982488 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa16f94f_be9c_4e2c_8dc3_c420db3ed639.slice/crio-f6706a204bf054a6bc0f63126cfdfb17c6a56c5963a6928ce1906dcf821a8d34 WatchSource:0}: Error finding container f6706a204bf054a6bc0f63126cfdfb17c6a56c5963a6928ce1906dcf821a8d34: Status 404 returned error can't find the container with id f6706a204bf054a6bc0f63126cfdfb17c6a56c5963a6928ce1906dcf821a8d34 Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.009805 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9"] Mar 20 11:15:00 crc kubenswrapper[4846]: W0320 11:15:00.020515 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7f86e9b_89bb_4558_aa0d_4bec37131052.slice/crio-5db53261e98cbc8007e862f48d0626c97c5e5d2bc712b5d60889e7cc17afa2f2 WatchSource:0}: Error finding container 5db53261e98cbc8007e862f48d0626c97c5e5d2bc712b5d60889e7cc17afa2f2: Status 404 returned error can't find the container with id 5db53261e98cbc8007e862f48d0626c97c5e5d2bc712b5d60889e7cc17afa2f2 Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.071558 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms"] Mar 20 11:15:00 crc kubenswrapper[4846]: W0320 11:15:00.085225 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod132090d4_b1d3_485e_9525_26fe1eecc448.slice/crio-9cc7f3a06f77f41dc93865f234a68e381efbc7fe5e44bdad59abad80931e3cba WatchSource:0}: Error finding container 9cc7f3a06f77f41dc93865f234a68e381efbc7fe5e44bdad59abad80931e3cba: Status 404 returned error can't find the container with id 9cc7f3a06f77f41dc93865f234a68e381efbc7fe5e44bdad59abad80931e3cba Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.088163 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp"] Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.089795 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dh2hx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-dr5ms_openstack-operators(132090d4-b1d3-485e-9525-26fe1eecc448): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.100986 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4tbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-884679f54-gqbrp_openstack-operators(64ce367c-20db-4598-8006-f51d8aa22bdd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.121675 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" podUID="132090d4-b1d3-485e-9525-26fe1eecc448" Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.121789 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" podUID="64ce367c-20db-4598-8006-f51d8aa22bdd" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.176773 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c"] Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.178552 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.181683 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.181714 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.228483 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c"] Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.242001 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j"] Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.365822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.365883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.366017 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.366050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.366070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgcvb\" (UniqueName: \"kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.366689 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.366753 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:02.366729487 +0000 UTC m=+1073.733137723 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.367213 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.367254 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:02.3672461 +0000 UTC m=+1073.733654336 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.433201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" event={"ID":"132090d4-b1d3-485e-9525-26fe1eecc448","Type":"ContainerStarted","Data":"9cc7f3a06f77f41dc93865f234a68e381efbc7fe5e44bdad59abad80931e3cba"} Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.435079 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" podUID="132090d4-b1d3-485e-9525-26fe1eecc448" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.436418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" event={"ID":"06eb056c-f3f0-4242-bbee-d201dfd37925","Type":"ContainerStarted","Data":"5f434b330658fb21de2108a46beae4a5030cf82a1e12b0151d70555edd1cad5d"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.446649 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" event={"ID":"9c7c760d-4ef7-41a0-bca6-c9b26d99fd52","Type":"ContainerStarted","Data":"86be8fc9e7ca196de6b07db719dd45646ca0316c91883f3aaa09ac02168d5534"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.454974 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerStarted","Data":"7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.459135 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" event={"ID":"0b70e378-a53e-4519-b415-64810a544874","Type":"ContainerStarted","Data":"1f32c956995659ff58b4ef9cb4105aed7aa914192c007a0a52257c0d41f6a8fb"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.460490 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" event={"ID":"2bba6588-37b3-453e-972e-393208ddd560","Type":"ContainerStarted","Data":"6ed68b19bf5b44a208ca0480b752d2959bdcbc7122cf8f9389a2f4bcc41ae078"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.462396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" event={"ID":"e7f86e9b-89bb-4558-aa0d-4bec37131052","Type":"ContainerStarted","Data":"5db53261e98cbc8007e862f48d0626c97c5e5d2bc712b5d60889e7cc17afa2f2"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.465304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" event={"ID":"64ce367c-20db-4598-8006-f51d8aa22bdd","Type":"ContainerStarted","Data":"81b7cedcaf479078a3efb30293230a48bb2032f45bde85dabde961e88d01a6f3"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.468118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.468162 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.468190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgcvb\" (UniqueName: \"kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: E0320 11:15:00.468207 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" podUID="64ce367c-20db-4598-8006-f51d8aa22bdd" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.469506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" event={"ID":"b155d1b3-b553-414b-9017-329ee45a9658","Type":"ContainerStarted","Data":"f00913083d74f1ff01f41f5430462d3b93a186af28d535485cbc40502bbc8b69"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.469735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.480313 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.495435 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgcvb\" (UniqueName: \"kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb\") pod \"collect-profiles-29566755-g7p9c\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.495792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" event={"ID":"86d13cbe-cc30-41b6-8829-1f5cbb99c48c","Type":"ContainerStarted","Data":"0a02ed5b7ea6b187cf494fc8bbfe0ffc55e5b1e97f5be15d74fc1a28b9670810"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.497816 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4622r" podStartSLOduration=3.620691083 podStartE2EDuration="7.497790767s" podCreationTimestamp="2026-03-20 11:14:53 +0000 UTC" firstStartedPulling="2026-03-20 11:14:55.170857918 +0000 UTC m=+1066.537266154" lastFinishedPulling="2026-03-20 11:14:59.047957602 +0000 UTC m=+1070.414365838" observedRunningTime="2026-03-20 11:15:00.491789154 +0000 UTC m=+1071.858197390" watchObservedRunningTime="2026-03-20 11:15:00.497790767 +0000 UTC m=+1071.864199003" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.500866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" event={"ID":"c88c55bf-7869-40e5-90f7-0bfdcb676b9e","Type":"ContainerStarted","Data":"a8d2e42d55719830fc531429818b6b9a08bf00b25aa12c2328caad0ab7b4f17b"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.503534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" event={"ID":"bd21bfeb-d2b2-4757-aa36-094c039951f0","Type":"ContainerStarted","Data":"bda49911acc94371fd727c427acd758a2bc1a53648d3cbf228b99134b1a1d26a"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.517502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" event={"ID":"fa16f94f-be9c-4e2c-8dc3-c420db3ed639","Type":"ContainerStarted","Data":"f6706a204bf054a6bc0f63126cfdfb17c6a56c5963a6928ce1906dcf821a8d34"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.525587 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.533624 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" event={"ID":"6be091f8-9109-4fce-a856-feacf04c76d4","Type":"ContainerStarted","Data":"c6f64aa94cc1adea70c2ecd6a1471ef0fa3f83119a866e346d006a2897697972"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.536366 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" event={"ID":"14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd","Type":"ContainerStarted","Data":"a58a42628874036dd493e29c8305b4f6dab95370f9fba4472cf7f50082aeea0f"} Mar 20 11:15:00 crc kubenswrapper[4846]: I0320 11:15:00.537786 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" event={"ID":"9bf60e8a-5221-44ba-8471-5b9e2692d2a7","Type":"ContainerStarted","Data":"62f16d9c06756da4eb9f30177f7caf671ab5f66668ce620d5a528b4602c3b231"} Mar 20 11:15:01 crc kubenswrapper[4846]: I0320 11:15:01.208172 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c"] Mar 20 11:15:01 crc kubenswrapper[4846]: W0320 11:15:01.230070 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc731722c_d7e1_4b4d_a590_26c27cb79ee8.slice/crio-1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562 WatchSource:0}: Error finding container 1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562: Status 404 returned error can't find the container with id 1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562 Mar 20 11:15:01 crc kubenswrapper[4846]: I0320 11:15:01.397181 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.397406 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.397527 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:05.397479536 +0000 UTC m=+1076.763887942 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:01 crc kubenswrapper[4846]: I0320 11:15:01.580700 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" event={"ID":"c731722c-d7e1-4b4d-a590-26c27cb79ee8","Type":"ContainerStarted","Data":"1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562"} Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.589152 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" podUID="64ce367c-20db-4598-8006-f51d8aa22bdd" Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.601568 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" podUID="132090d4-b1d3-485e-9525-26fe1eecc448" Mar 20 11:15:01 crc kubenswrapper[4846]: I0320 11:15:01.907142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.907333 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:01 crc kubenswrapper[4846]: E0320 11:15:01.907428 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:05.907403971 +0000 UTC m=+1077.273812207 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:02 crc kubenswrapper[4846]: I0320 11:15:02.428056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:02 crc kubenswrapper[4846]: E0320 11:15:02.428749 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:15:02 crc kubenswrapper[4846]: E0320 11:15:02.429612 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:06.429572097 +0000 UTC m=+1077.795980513 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:15:02 crc kubenswrapper[4846]: I0320 11:15:02.431164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:02 crc kubenswrapper[4846]: E0320 11:15:02.431345 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:15:02 crc kubenswrapper[4846]: E0320 11:15:02.431382 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:06.4313726 +0000 UTC m=+1077.797780836 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:15:02 crc kubenswrapper[4846]: I0320 11:15:02.615448 4846 generic.go:334] "Generic (PLEG): container finished" podID="c731722c-d7e1-4b4d-a590-26c27cb79ee8" containerID="7eb95debfbfa0adc05fb81b581561541e2fbcd022bde093c05d3ca847835a8fb" exitCode=0 Mar 20 11:15:02 crc kubenswrapper[4846]: I0320 11:15:02.615515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" event={"ID":"c731722c-d7e1-4b4d-a590-26c27cb79ee8","Type":"ContainerDied","Data":"7eb95debfbfa0adc05fb81b581561541e2fbcd022bde093c05d3ca847835a8fb"} Mar 20 11:15:03 crc kubenswrapper[4846]: I0320 11:15:03.677486 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:03 crc kubenswrapper[4846]: I0320 11:15:03.677564 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:03 crc kubenswrapper[4846]: I0320 11:15:03.915801 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.046721 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.092821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume\") pod \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.092954 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgcvb\" (UniqueName: \"kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb\") pod \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.093028 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume\") pod \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\" (UID: \"c731722c-d7e1-4b4d-a590-26c27cb79ee8\") " Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.093972 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume" (OuterVolumeSpecName: "config-volume") pod "c731722c-d7e1-4b4d-a590-26c27cb79ee8" (UID: "c731722c-d7e1-4b4d-a590-26c27cb79ee8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.102638 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c731722c-d7e1-4b4d-a590-26c27cb79ee8" (UID: "c731722c-d7e1-4b4d-a590-26c27cb79ee8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.103335 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb" (OuterVolumeSpecName: "kube-api-access-xgcvb") pod "c731722c-d7e1-4b4d-a590-26c27cb79ee8" (UID: "c731722c-d7e1-4b4d-a590-26c27cb79ee8"). InnerVolumeSpecName "kube-api-access-xgcvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.195411 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c731722c-d7e1-4b4d-a590-26c27cb79ee8-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.195446 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgcvb\" (UniqueName: \"kubernetes.io/projected/c731722c-d7e1-4b4d-a590-26c27cb79ee8-kube-api-access-xgcvb\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.195455 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c731722c-d7e1-4b4d-a590-26c27cb79ee8-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.688796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" event={"ID":"c731722c-d7e1-4b4d-a590-26c27cb79ee8","Type":"ContainerDied","Data":"1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562"} Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.688885 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1555564fd4478f94a8dd71e006472045b3e58b64e9565e9ee5c2f62828073562" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.688819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566755-g7p9c" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.765123 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:04 crc kubenswrapper[4846]: I0320 11:15:04.825583 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:15:05 crc kubenswrapper[4846]: I0320 11:15:05.438785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:05 crc kubenswrapper[4846]: E0320 11:15:05.438995 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:05 crc kubenswrapper[4846]: E0320 11:15:05.439097 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:13.439070517 +0000 UTC m=+1084.805478753 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:05 crc kubenswrapper[4846]: I0320 11:15:05.949948 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:05 crc kubenswrapper[4846]: E0320 11:15:05.950138 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:05 crc kubenswrapper[4846]: E0320 11:15:05.950203 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:13.95018439 +0000 UTC m=+1085.316592626 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.462584 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.462664 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:06 crc kubenswrapper[4846]: E0320 11:15:06.462922 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:15:06 crc kubenswrapper[4846]: E0320 11:15:06.462989 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:14.462967362 +0000 UTC m=+1085.829375598 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:15:06 crc kubenswrapper[4846]: E0320 11:15:06.463483 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:15:06 crc kubenswrapper[4846]: E0320 11:15:06.463512 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:14.463503425 +0000 UTC m=+1085.829911661 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.573851 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:06 crc kubenswrapper[4846]: E0320 11:15:06.574495 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c731722c-d7e1-4b4d-a590-26c27cb79ee8" containerName="collect-profiles" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.574569 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c731722c-d7e1-4b4d-a590-26c27cb79ee8" containerName="collect-profiles" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.575082 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c731722c-d7e1-4b4d-a590-26c27cb79ee8" containerName="collect-profiles" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.576202 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.606815 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.666763 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.667192 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.667347 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdrbj\" (UniqueName: \"kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.702828 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4622r" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" containerID="cri-o://7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" gracePeriod=2 Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.769418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.769494 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.769572 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdrbj\" (UniqueName: \"kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.770357 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.770500 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.791431 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdrbj\" (UniqueName: \"kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj\") pod \"community-operators-ms9js\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:06 crc kubenswrapper[4846]: I0320 11:15:06.902601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:07 crc kubenswrapper[4846]: I0320 11:15:07.715302 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" exitCode=0 Mar 20 11:15:07 crc kubenswrapper[4846]: I0320 11:15:07.715355 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerDied","Data":"7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7"} Mar 20 11:15:13 crc kubenswrapper[4846]: I0320 11:15:13.481393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.482516 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.482589 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert podName:1132b3dc-feeb-41ce-a579-d5b089b86597 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:29.482566012 +0000 UTC m=+1100.848974238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert") pod "infra-operator-controller-manager-594f6f699b-vmg46" (UID: "1132b3dc-feeb-41ce-a579-d5b089b86597") : secret "infra-operator-webhook-server-cert" not found Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.674765 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.675525 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.675708 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.675738 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-4622r" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" Mar 20 11:15:13 crc kubenswrapper[4846]: I0320 11:15:13.990620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.990871 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:13 crc kubenswrapper[4846]: E0320 11:15:13.991023 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert podName:7d330820-2aac-4076-a205-8dedd331cae1 nodeName:}" failed. No retries permitted until 2026-03-20 11:15:29.99099072 +0000 UTC m=+1101.357398956 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-46rnd" (UID: "7d330820-2aac-4076-a205-8dedd331cae1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.271879 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900" Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.272109 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lpvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-67dd5f86f5-thhs6_openstack-operators(24727b45-3896-4b9f-a91e-62434aa32ac7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.273769 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" podUID="24727b45-3896-4b9f-a91e-62434aa32ac7" Mar 20 11:15:14 crc kubenswrapper[4846]: I0320 11:15:14.500582 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:14 crc kubenswrapper[4846]: I0320 11:15:14.500694 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.500954 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.501038 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.501140 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:30.501108948 +0000 UTC m=+1101.867517344 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "webhook-server-cert" not found Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.501221 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs podName:de215fcd-2be6-40ca-8ae1-d628b6b4280b nodeName:}" failed. No retries permitted until 2026-03-20 11:15:30.50119582 +0000 UTC m=+1101.867604056 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs") pod "openstack-operator-controller-manager-7575fb585b-v27cw" (UID: "de215fcd-2be6-40ca-8ae1-d628b6b4280b") : secret "metrics-server-cert" not found Mar 20 11:15:14 crc kubenswrapper[4846]: E0320 11:15:14.785459 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900\\\"\"" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" podUID="24727b45-3896-4b9f-a91e-62434aa32ac7" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.098097 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.098361 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lthg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-ppjfc_openstack-operators(14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.099440 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" podUID="14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.793852 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" podUID="14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.860109 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.860408 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lsf77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6f787dddc9-g8hzn_openstack-operators(c88c55bf-7869-40e5-90f7-0bfdcb676b9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:15 crc kubenswrapper[4846]: E0320 11:15:15.861609 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" podUID="c88c55bf-7869-40e5-90f7-0bfdcb676b9e" Mar 20 11:15:16 crc kubenswrapper[4846]: E0320 11:15:16.505780 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d" Mar 20 11:15:16 crc kubenswrapper[4846]: E0320 11:15:16.506046 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5jc87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-79df6bcc97-4m4lg_openstack-operators(380b5a6a-c151-4763-a86b-c8e53b6f1aa2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:16 crc kubenswrapper[4846]: E0320 11:15:16.507963 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" podUID="380b5a6a-c151-4763-a86b-c8e53b6f1aa2" Mar 20 11:15:16 crc kubenswrapper[4846]: E0320 11:15:16.804134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" podUID="c88c55bf-7869-40e5-90f7-0bfdcb676b9e" Mar 20 11:15:16 crc kubenswrapper[4846]: E0320 11:15:16.804401 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d\\\"\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" podUID="380b5a6a-c151-4763-a86b-c8e53b6f1aa2" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.148650 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.149328 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vjbhm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-dvj5j_openstack-operators(b155d1b3-b553-414b-9017-329ee45a9658): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.150622 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" podUID="b155d1b3-b553-414b-9017-329ee45a9658" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.817630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" podUID="b155d1b3-b553-414b-9017-329ee45a9658" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.949463 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.949733 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2hdgj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-588d4d986b-ht246_openstack-operators(9bf60e8a-5221-44ba-8471-5b9e2692d2a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:17 crc kubenswrapper[4846]: E0320 11:15:17.951021 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" podUID="9bf60e8a-5221-44ba-8471-5b9e2692d2a7" Mar 20 11:15:18 crc kubenswrapper[4846]: E0320 11:15:18.741233 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" Mar 20 11:15:18 crc kubenswrapper[4846]: E0320 11:15:18.741462 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mf72p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-b8ng9_openstack-operators(e7f86e9b-89bb-4558-aa0d-4bec37131052): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:18 crc kubenswrapper[4846]: E0320 11:15:18.743335 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" podUID="e7f86e9b-89bb-4558-aa0d-4bec37131052" Mar 20 11:15:18 crc kubenswrapper[4846]: E0320 11:15:18.825835 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" podUID="e7f86e9b-89bb-4558-aa0d-4bec37131052" Mar 20 11:15:18 crc kubenswrapper[4846]: E0320 11:15:18.828836 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad\\\"\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" podUID="9bf60e8a-5221-44ba-8471-5b9e2692d2a7" Mar 20 11:15:19 crc kubenswrapper[4846]: E0320 11:15:19.496738 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a" Mar 20 11:15:19 crc kubenswrapper[4846]: E0320 11:15:19.496980 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qxj26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-767865f676-9wzfn_openstack-operators(fa16f94f-be9c-4e2c-8dc3-c420db3ed639): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:19 crc kubenswrapper[4846]: E0320 11:15:19.498175 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" podUID="fa16f94f-be9c-4e2c-8dc3-c420db3ed639" Mar 20 11:15:19 crc kubenswrapper[4846]: E0320 11:15:19.832617 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" podUID="fa16f94f-be9c-4e2c-8dc3-c420db3ed639" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.155646 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.155908 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dkfsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5b9f45d989-rrx9r_openstack-operators(06eb056c-f3f0-4242-bbee-d201dfd37925): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.157099 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" podUID="06eb056c-f3f0-4242-bbee-d201dfd37925" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.792022 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.792334 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kkb6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-55f864c847-hgqt9_openstack-operators(86d13cbe-cc30-41b6-8829-1f5cbb99c48c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.793595 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" podUID="86d13cbe-cc30-41b6-8829-1f5cbb99c48c" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.842497 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" podUID="06eb056c-f3f0-4242-bbee-d201dfd37925" Mar 20 11:15:20 crc kubenswrapper[4846]: E0320 11:15:20.842958 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" podUID="86d13cbe-cc30-41b6-8829-1f5cbb99c48c" Mar 20 11:15:21 crc kubenswrapper[4846]: E0320 11:15:21.531701 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113" Mar 20 11:15:21 crc kubenswrapper[4846]: E0320 11:15:21.532281 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r8gzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-8464cc45fb-qg5jf_openstack-operators(6be091f8-9109-4fce-a856-feacf04c76d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:21 crc kubenswrapper[4846]: E0320 11:15:21.533973 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" podUID="6be091f8-9109-4fce-a856-feacf04c76d4" Mar 20 11:15:21 crc kubenswrapper[4846]: E0320 11:15:21.846718 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" podUID="6be091f8-9109-4fce-a856-feacf04c76d4" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.321979 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.322326 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxhv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-f45j6_openstack-operators(bd21bfeb-d2b2-4757-aa36-094c039951f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.323674 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" podUID="bd21bfeb-d2b2-4757-aa36-094c039951f0" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.771136 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.771651 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qn9ck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-qwptc_openstack-operators(2bba6588-37b3-453e-972e-393208ddd560): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.773003 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" podUID="2bba6588-37b3-453e-972e-393208ddd560" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.852855 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" podUID="bd21bfeb-d2b2-4757-aa36-094c039951f0" Mar 20 11:15:22 crc kubenswrapper[4846]: E0320 11:15:22.853160 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" podUID="2bba6588-37b3-453e-972e-393208ddd560" Mar 20 11:15:23 crc kubenswrapper[4846]: E0320 11:15:23.673686 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:23 crc kubenswrapper[4846]: E0320 11:15:23.674169 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:23 crc kubenswrapper[4846]: E0320 11:15:23.674528 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" cmd=["grpc_health_probe","-addr=:50051"] Mar 20 11:15:23 crc kubenswrapper[4846]: E0320 11:15:23.674572 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-4622r" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" Mar 20 11:15:23 crc kubenswrapper[4846]: I0320 11:15:23.872277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4622r" event={"ID":"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191","Type":"ContainerDied","Data":"6f9eb2b36e70fb6093481b027d62b010f716d9bd9d170d323cdefa93aee264b5"} Mar 20 11:15:23 crc kubenswrapper[4846]: I0320 11:15:23.872630 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f9eb2b36e70fb6093481b027d62b010f716d9bd9d170d323cdefa93aee264b5" Mar 20 11:15:23 crc kubenswrapper[4846]: I0320 11:15:23.959093 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.063694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content\") pod \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.064284 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities\") pod \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.064336 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkcpd\" (UniqueName: \"kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd\") pod \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\" (UID: \"ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191\") " Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.066718 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities" (OuterVolumeSpecName: "utilities") pod "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" (UID: "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.074322 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd" (OuterVolumeSpecName: "kube-api-access-hkcpd") pod "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" (UID: "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191"). InnerVolumeSpecName "kube-api-access-hkcpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.129001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" (UID: "ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.165749 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.165811 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.165821 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkcpd\" (UniqueName: \"kubernetes.io/projected/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191-kube-api-access-hkcpd\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.305354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:24 crc kubenswrapper[4846]: W0320 11:15:24.311458 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4883575_f600_4012_ac6d_47dfdf2a1394.slice/crio-011d5b0e8d7b0770eeb885099ab305668a2e82ada46a7eb972f1fc659114af8c WatchSource:0}: Error finding container 011d5b0e8d7b0770eeb885099ab305668a2e82ada46a7eb972f1fc659114af8c: Status 404 returned error can't find the container with id 011d5b0e8d7b0770eeb885099ab305668a2e82ada46a7eb972f1fc659114af8c Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.881627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" event={"ID":"aff6e349-68ea-4f3c-8514-b2a7c7ffed23","Type":"ContainerStarted","Data":"551c181185706f9baf71c80dbbd42dc4e9a19405970e44991c40f14a36dc03b2"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.883064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" event={"ID":"0b70e378-a53e-4519-b415-64810a544874","Type":"ContainerStarted","Data":"74c5b1c1b7a0ed073ae6bbb31c079cd4b78a7d82a08bcd9b58d7f4e75e44c152"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.883228 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.884642 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerID="6e3b3becea99460e3829656e31ccd241d0516848dc5bcd7ee238ac7883745cba" exitCode=0 Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.884736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerDied","Data":"6e3b3becea99460e3829656e31ccd241d0516848dc5bcd7ee238ac7883745cba"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.884792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerStarted","Data":"011d5b0e8d7b0770eeb885099ab305668a2e82ada46a7eb972f1fc659114af8c"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.887205 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" event={"ID":"9c7c760d-4ef7-41a0-bca6-c9b26d99fd52","Type":"ContainerStarted","Data":"52fb0ef204016dff13788feed0b5f38da6e804e0d7be161260b8040699e74b58"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.887348 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.892217 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" event={"ID":"132090d4-b1d3-485e-9525-26fe1eecc448","Type":"ContainerStarted","Data":"628bd5f6685c77d6c0e1b26d761c7a26a9547a89ecd019dc3d0d42a81ad7c06e"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.892505 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.897770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" event={"ID":"64ce367c-20db-4598-8006-f51d8aa22bdd","Type":"ContainerStarted","Data":"c1d9d70aba17f010f1a20f494af7c5023488f9173b5ad3a57cedd9848d6cbbca"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.898242 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.899968 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" event={"ID":"529816a9-1274-4a4f-88b2-6e0b426075a0","Type":"ContainerStarted","Data":"dd6357df7c49739386786102e90a849d5ab504dea634acad1828507a30acf871"} Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.900030 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4622r" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.911152 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" podStartSLOduration=3.147713368 podStartE2EDuration="27.911130184s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.034089871 +0000 UTC m=+1070.400498107" lastFinishedPulling="2026-03-20 11:15:23.797506687 +0000 UTC m=+1095.163914923" observedRunningTime="2026-03-20 11:15:24.908732737 +0000 UTC m=+1096.275140993" watchObservedRunningTime="2026-03-20 11:15:24.911130184 +0000 UTC m=+1096.277538420" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.936442 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" podStartSLOduration=4.091409479 podStartE2EDuration="27.936417988s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.953045641 +0000 UTC m=+1071.319453877" lastFinishedPulling="2026-03-20 11:15:23.79805411 +0000 UTC m=+1095.164462386" observedRunningTime="2026-03-20 11:15:24.930591529 +0000 UTC m=+1096.296999765" watchObservedRunningTime="2026-03-20 11:15:24.936417988 +0000 UTC m=+1096.302826224" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.952332 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" podStartSLOduration=3.914422454 podStartE2EDuration="27.952309518s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.760399261 +0000 UTC m=+1071.126807507" lastFinishedPulling="2026-03-20 11:15:23.798286325 +0000 UTC m=+1095.164694571" observedRunningTime="2026-03-20 11:15:24.949827158 +0000 UTC m=+1096.316235394" watchObservedRunningTime="2026-03-20 11:15:24.952309518 +0000 UTC m=+1096.318717764" Mar 20 11:15:24 crc kubenswrapper[4846]: I0320 11:15:24.966043 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" podStartSLOduration=3.880288558 podStartE2EDuration="27.966015324s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:58.640760541 +0000 UTC m=+1070.007168777" lastFinishedPulling="2026-03-20 11:15:22.726487317 +0000 UTC m=+1094.092895543" observedRunningTime="2026-03-20 11:15:24.963090625 +0000 UTC m=+1096.329498861" watchObservedRunningTime="2026-03-20 11:15:24.966015324 +0000 UTC m=+1096.332423560" Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.027509 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" podStartSLOduration=4.275285208 podStartE2EDuration="28.027484242s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:00.100762367 +0000 UTC m=+1071.467170603" lastFinishedPulling="2026-03-20 11:15:23.852961401 +0000 UTC m=+1095.219369637" observedRunningTime="2026-03-20 11:15:25.023703522 +0000 UTC m=+1096.390111778" watchObservedRunningTime="2026-03-20 11:15:25.027484242 +0000 UTC m=+1096.393892478" Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.043324 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" podStartSLOduration=4.276913807 podStartE2EDuration="28.04329336s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:00.089626431 +0000 UTC m=+1071.456034667" lastFinishedPulling="2026-03-20 11:15:23.856005954 +0000 UTC m=+1095.222414220" observedRunningTime="2026-03-20 11:15:25.037748248 +0000 UTC m=+1096.404156484" watchObservedRunningTime="2026-03-20 11:15:25.04329336 +0000 UTC m=+1096.409701596" Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.060261 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.066675 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4622r"] Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.332301 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" path="/var/lib/kubelet/pods/ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191/volumes" Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.908229 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerStarted","Data":"977b19b8b56784c5c72bc1a4dec37ac96742a737204fa238d9dd1701059afe13"} Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.910367 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:15:25 crc kubenswrapper[4846]: I0320 11:15:25.910431 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:15:26 crc kubenswrapper[4846]: I0320 11:15:26.919008 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerID="977b19b8b56784c5c72bc1a4dec37ac96742a737204fa238d9dd1701059afe13" exitCode=0 Mar 20 11:15:26 crc kubenswrapper[4846]: I0320 11:15:26.919164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerDied","Data":"977b19b8b56784c5c72bc1a4dec37ac96742a737204fa238d9dd1701059afe13"} Mar 20 11:15:27 crc kubenswrapper[4846]: I0320 11:15:27.931936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerStarted","Data":"c4920aaef9fc2489b79d885c1945b573a0cebfed39a9ceb6f659750e966456ab"} Mar 20 11:15:27 crc kubenswrapper[4846]: I0320 11:15:27.953361 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ms9js" podStartSLOduration=19.492322591 podStartE2EDuration="21.953341696s" podCreationTimestamp="2026-03-20 11:15:06 +0000 UTC" firstStartedPulling="2026-03-20 11:15:24.886761043 +0000 UTC m=+1096.253169279" lastFinishedPulling="2026-03-20 11:15:27.347780138 +0000 UTC m=+1098.714188384" observedRunningTime="2026-03-20 11:15:27.951074782 +0000 UTC m=+1099.317483028" watchObservedRunningTime="2026-03-20 11:15:27.953341696 +0000 UTC m=+1099.319749932" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.559456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.566387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1132b3dc-feeb-41ce-a579-d5b089b86597-cert\") pod \"infra-operator-controller-manager-594f6f699b-vmg46\" (UID: \"1132b3dc-feeb-41ce-a579-d5b089b86597\") " pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.769063 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r5shd" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.776387 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.953619 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" event={"ID":"14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd","Type":"ContainerStarted","Data":"ba9576a333a31e3cb591c08dffbc6059a978179312289476042e02bf0eaf0463"} Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.955130 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:15:29 crc kubenswrapper[4846]: I0320 11:15:29.979986 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" podStartSLOduration=4.190111559 podStartE2EDuration="32.979961114s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.983699983 +0000 UTC m=+1071.350108219" lastFinishedPulling="2026-03-20 11:15:28.773549528 +0000 UTC m=+1100.139957774" observedRunningTime="2026-03-20 11:15:29.976252371 +0000 UTC m=+1101.342660617" watchObservedRunningTime="2026-03-20 11:15:29.979961114 +0000 UTC m=+1101.346369370" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.061637 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46"] Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.066671 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.084955 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d330820-2aac-4076-a205-8dedd331cae1-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-46rnd\" (UID: \"7d330820-2aac-4076-a205-8dedd331cae1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.313216 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cvb79" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.321921 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.582114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.582506 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.587363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-webhook-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.595297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/de215fcd-2be6-40ca-8ae1-d628b6b4280b-metrics-certs\") pod \"openstack-operator-controller-manager-7575fb585b-v27cw\" (UID: \"de215fcd-2be6-40ca-8ae1-d628b6b4280b\") " pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.656320 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd"] Mar 20 11:15:30 crc kubenswrapper[4846]: W0320 11:15:30.660981 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d330820_2aac_4076_a205_8dedd331cae1.slice/crio-cb2d22faf6daf4ad0b28ad2563e05cefea8e5e1179988868be7d375299895ff9 WatchSource:0}: Error finding container cb2d22faf6daf4ad0b28ad2563e05cefea8e5e1179988868be7d375299895ff9: Status 404 returned error can't find the container with id cb2d22faf6daf4ad0b28ad2563e05cefea8e5e1179988868be7d375299895ff9 Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.898977 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sxrj5" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.907188 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.963038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" event={"ID":"c88c55bf-7869-40e5-90f7-0bfdcb676b9e","Type":"ContainerStarted","Data":"1cdf0eb45ab56d3b514806fc7b95fc58c1b942438b959456e8616a776a239ad8"} Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.964343 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.988628 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" podStartSLOduration=3.389138622 podStartE2EDuration="33.988606278s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.702282464 +0000 UTC m=+1071.068690690" lastFinishedPulling="2026-03-20 11:15:30.30175011 +0000 UTC m=+1101.668158346" observedRunningTime="2026-03-20 11:15:30.980366356 +0000 UTC m=+1102.346774592" watchObservedRunningTime="2026-03-20 11:15:30.988606278 +0000 UTC m=+1102.355014514" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.992014 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" event={"ID":"1132b3dc-feeb-41ce-a579-d5b089b86597","Type":"ContainerStarted","Data":"4ff9b9e4fbab59de704aa10d02dbc39a08fa6fc8580b261765465a22a666b256"} Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.997434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" event={"ID":"24727b45-3896-4b9f-a91e-62434aa32ac7","Type":"ContainerStarted","Data":"188cc6028d07b96e6cbac45b0e7dccf7eaded99b6a23c2028641602a49e24590"} Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.997829 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:15:30 crc kubenswrapper[4846]: I0320 11:15:30.999428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" event={"ID":"7d330820-2aac-4076-a205-8dedd331cae1","Type":"ContainerStarted","Data":"cb2d22faf6daf4ad0b28ad2563e05cefea8e5e1179988868be7d375299895ff9"} Mar 20 11:15:31 crc kubenswrapper[4846]: I0320 11:15:31.023574 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" podStartSLOduration=3.174834964 podStartE2EDuration="34.023552069s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.121846326 +0000 UTC m=+1070.488254572" lastFinishedPulling="2026-03-20 11:15:29.970563431 +0000 UTC m=+1101.336971677" observedRunningTime="2026-03-20 11:15:31.022097518 +0000 UTC m=+1102.388505744" watchObservedRunningTime="2026-03-20 11:15:31.023552069 +0000 UTC m=+1102.389960305" Mar 20 11:15:31 crc kubenswrapper[4846]: I0320 11:15:31.408814 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw"] Mar 20 11:15:32 crc kubenswrapper[4846]: I0320 11:15:32.009215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" event={"ID":"380b5a6a-c151-4763-a86b-c8e53b6f1aa2","Type":"ContainerStarted","Data":"808c0db53c58fa9b4a943c624e15c945a4ebb9d77f8b1926baf6943b8e27990b"} Mar 20 11:15:32 crc kubenswrapper[4846]: I0320 11:15:32.010281 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:15:32 crc kubenswrapper[4846]: I0320 11:15:32.036702 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" podStartSLOduration=3.124415121 podStartE2EDuration="35.036679368s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:58.949989013 +0000 UTC m=+1070.316397249" lastFinishedPulling="2026-03-20 11:15:30.86225326 +0000 UTC m=+1102.228661496" observedRunningTime="2026-03-20 11:15:32.031515974 +0000 UTC m=+1103.397924220" watchObservedRunningTime="2026-03-20 11:15:32.036679368 +0000 UTC m=+1103.403087604" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.075499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" event={"ID":"6be091f8-9109-4fce-a856-feacf04c76d4","Type":"ContainerStarted","Data":"871b3e8c4567b4a4ffbb1f969bdf0825a28c4e6c8b15d2e21eac225a921adff6"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.076393 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.078590 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" event={"ID":"b155d1b3-b553-414b-9017-329ee45a9658","Type":"ContainerStarted","Data":"d8e5e6329e9ebd18de93de4ea78cd6582096007312986767afea6aacccab3560"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.078858 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.081414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" event={"ID":"06eb056c-f3f0-4242-bbee-d201dfd37925","Type":"ContainerStarted","Data":"8ba70a51bfe734b6f7c6a5dd8b7bab995ab967e7f513124d67033e0a933e6f64"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.081923 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.086045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" event={"ID":"7d330820-2aac-4076-a205-8dedd331cae1","Type":"ContainerStarted","Data":"684667d444fa436274ac1afa9fcdbf7dc6e931b1b2463a8d8bdeac9f1dc6f13d"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.086246 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.087613 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" event={"ID":"e7f86e9b-89bb-4558-aa0d-4bec37131052","Type":"ContainerStarted","Data":"85dcbe8922df5bce0c7077297f5b877ce21859a8a8fcf955ea1dd3d1e717c548"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.088124 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.090453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" event={"ID":"fa16f94f-be9c-4e2c-8dc3-c420db3ed639","Type":"ContainerStarted","Data":"1bccfbeffb3f039108cb37df0d97e97b69995a34306ab5b8dd181d10265ef2ab"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.091128 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.108686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" event={"ID":"9bf60e8a-5221-44ba-8471-5b9e2692d2a7","Type":"ContainerStarted","Data":"c91ee616a08bfa4dc4cd552b5c063d4c47d44179ad526238d2f7db7ce44e851b"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.109653 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.113197 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" event={"ID":"de215fcd-2be6-40ca-8ae1-d628b6b4280b","Type":"ContainerStarted","Data":"2ca9831f896b494d584796f8b880caa614cdc9ddc7c3181c9ce25ee6ab434622"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.113257 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" event={"ID":"de215fcd-2be6-40ca-8ae1-d628b6b4280b","Type":"ContainerStarted","Data":"138f00a3243644ae7a2a3f671ff16fbcf1dc2bcc67a42f43542b670b128dad11"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.114075 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.116136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" event={"ID":"1132b3dc-feeb-41ce-a579-d5b089b86597","Type":"ContainerStarted","Data":"b7c8cc5995a9f848af877a47d156bc353a8f4c932e603d1c1686d36145b94eb4"} Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.116660 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.150617 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" podStartSLOduration=3.458626473 podStartE2EDuration="37.150594908s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.552217471 +0000 UTC m=+1070.918625707" lastFinishedPulling="2026-03-20 11:15:33.244185896 +0000 UTC m=+1104.610594142" observedRunningTime="2026-03-20 11:15:34.121068629 +0000 UTC m=+1105.487476865" watchObservedRunningTime="2026-03-20 11:15:34.150594908 +0000 UTC m=+1105.517003144" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.220031 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" podStartSLOduration=3.4668359349999998 podStartE2EDuration="37.220007787s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.528108045 +0000 UTC m=+1070.894516281" lastFinishedPulling="2026-03-20 11:15:33.281279897 +0000 UTC m=+1104.647688133" observedRunningTime="2026-03-20 11:15:34.155884487 +0000 UTC m=+1105.522292733" watchObservedRunningTime="2026-03-20 11:15:34.220007787 +0000 UTC m=+1105.586416033" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.222419 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" podStartSLOduration=34.047068359 podStartE2EDuration="37.222411764s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:30.071053772 +0000 UTC m=+1101.437461998" lastFinishedPulling="2026-03-20 11:15:33.246397167 +0000 UTC m=+1104.612805403" observedRunningTime="2026-03-20 11:15:34.219476622 +0000 UTC m=+1105.585884868" watchObservedRunningTime="2026-03-20 11:15:34.222411764 +0000 UTC m=+1105.588820000" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.260681 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" podStartSLOduration=3.944212057 podStartE2EDuration="37.260659209s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.959384362 +0000 UTC m=+1071.325792598" lastFinishedPulling="2026-03-20 11:15:33.275831494 +0000 UTC m=+1104.642239750" observedRunningTime="2026-03-20 11:15:34.252913511 +0000 UTC m=+1105.619321747" watchObservedRunningTime="2026-03-20 11:15:34.260659209 +0000 UTC m=+1105.627067445" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.303544 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" podStartSLOduration=4.013353441 podStartE2EDuration="37.303515582s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.992027002 +0000 UTC m=+1071.358435238" lastFinishedPulling="2026-03-20 11:15:33.282189143 +0000 UTC m=+1104.648597379" observedRunningTime="2026-03-20 11:15:34.298941214 +0000 UTC m=+1105.665349470" watchObservedRunningTime="2026-03-20 11:15:34.303515582 +0000 UTC m=+1105.669923818" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.385576 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" podStartSLOduration=34.776619586 podStartE2EDuration="37.385552446s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:30.663404986 +0000 UTC m=+1102.029813222" lastFinishedPulling="2026-03-20 11:15:33.272337846 +0000 UTC m=+1104.638746082" observedRunningTime="2026-03-20 11:15:34.374865935 +0000 UTC m=+1105.741274171" watchObservedRunningTime="2026-03-20 11:15:34.385552446 +0000 UTC m=+1105.751960682" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.436538 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" podStartSLOduration=4.220985084 podStartE2EDuration="37.436514376s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:00.062648658 +0000 UTC m=+1071.429056894" lastFinishedPulling="2026-03-20 11:15:33.27817794 +0000 UTC m=+1104.644586186" observedRunningTime="2026-03-20 11:15:34.433830971 +0000 UTC m=+1105.800239207" watchObservedRunningTime="2026-03-20 11:15:34.436514376 +0000 UTC m=+1105.802922602" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.490320 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" podStartSLOduration=36.490301748 podStartE2EDuration="36.490301748s" podCreationTimestamp="2026-03-20 11:14:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:15:34.484077352 +0000 UTC m=+1105.850485588" watchObservedRunningTime="2026-03-20 11:15:34.490301748 +0000 UTC m=+1105.856709984" Mar 20 11:15:34 crc kubenswrapper[4846]: I0320 11:15:34.523457 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" podStartSLOduration=4.496451115 podStartE2EDuration="37.523427017s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:15:00.251371313 +0000 UTC m=+1071.617779549" lastFinishedPulling="2026-03-20 11:15:33.278347215 +0000 UTC m=+1104.644755451" observedRunningTime="2026-03-20 11:15:34.519040644 +0000 UTC m=+1105.885448880" watchObservedRunningTime="2026-03-20 11:15:34.523427017 +0000 UTC m=+1105.889835253" Mar 20 11:15:35 crc kubenswrapper[4846]: I0320 11:15:35.127804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" event={"ID":"86d13cbe-cc30-41b6-8829-1f5cbb99c48c","Type":"ContainerStarted","Data":"f8da59da5c212c115eff78fbf16b4f8b633a1b81eada91b535e4fd84b8dca253"} Mar 20 11:15:35 crc kubenswrapper[4846]: I0320 11:15:35.146024 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" podStartSLOduration=3.008064701 podStartE2EDuration="38.146006089s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.704304152 +0000 UTC m=+1071.070712388" lastFinishedPulling="2026-03-20 11:15:34.84224554 +0000 UTC m=+1106.208653776" observedRunningTime="2026-03-20 11:15:35.141708669 +0000 UTC m=+1106.508116935" watchObservedRunningTime="2026-03-20 11:15:35.146006089 +0000 UTC m=+1106.512414325" Mar 20 11:15:36 crc kubenswrapper[4846]: I0320 11:15:36.137043 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" event={"ID":"2bba6588-37b3-453e-972e-393208ddd560","Type":"ContainerStarted","Data":"74647af7716903387f621779b826f0676ef7c429c4a4f31a6ae45367cfc7e596"} Mar 20 11:15:36 crc kubenswrapper[4846]: I0320 11:15:36.903071 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:36 crc kubenswrapper[4846]: I0320 11:15:36.903464 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:36 crc kubenswrapper[4846]: I0320 11:15:36.968617 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.203119 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.255003 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.733826 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zncs8" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.761777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-ggrpk" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.799999 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-4m4lg" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.842112 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-thhs6" Mar 20 11:15:37 crc kubenswrapper[4846]: I0320 11:15:37.982704 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-g8hzn" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.171653 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.190459 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-r7pgd" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.413675 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-9wzfn" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.478515 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-rrx9r" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.600176 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-gqbrp" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.660396 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-dr5ms" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.691599 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-4fww7" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.827539 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-ppjfc" Mar 20 11:15:38 crc kubenswrapper[4846]: I0320 11:15:38.986494 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-b8ng9" Mar 20 11:15:39 crc kubenswrapper[4846]: I0320 11:15:39.069720 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-dvj5j" Mar 20 11:15:39 crc kubenswrapper[4846]: I0320 11:15:39.157712 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ms9js" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="registry-server" containerID="cri-o://c4920aaef9fc2489b79d885c1945b573a0cebfed39a9ceb6f659750e966456ab" gracePeriod=2 Mar 20 11:15:39 crc kubenswrapper[4846]: I0320 11:15:39.790854 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-594f6f699b-vmg46" Mar 20 11:15:40 crc kubenswrapper[4846]: I0320 11:15:40.330145 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-46rnd" Mar 20 11:15:40 crc kubenswrapper[4846]: I0320 11:15:40.917346 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7575fb585b-v27cw" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.196061 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerID="c4920aaef9fc2489b79d885c1945b573a0cebfed39a9ceb6f659750e966456ab" exitCode=0 Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.196302 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerDied","Data":"c4920aaef9fc2489b79d885c1945b573a0cebfed39a9ceb6f659750e966456ab"} Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.196670 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.199051 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.217245 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-qwptc" podStartSLOduration=10.256726811 podStartE2EDuration="46.217219704s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.760681448 +0000 UTC m=+1071.127089694" lastFinishedPulling="2026-03-20 11:15:35.721174351 +0000 UTC m=+1107.087582587" observedRunningTime="2026-03-20 11:15:43.21318639 +0000 UTC m=+1114.579594626" watchObservedRunningTime="2026-03-20 11:15:43.217219704 +0000 UTC m=+1114.583627940" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.578314 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.731684 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities\") pod \"f4883575-f600-4012-ac6d-47dfdf2a1394\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.732147 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content\") pod \"f4883575-f600-4012-ac6d-47dfdf2a1394\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.732328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdrbj\" (UniqueName: \"kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj\") pod \"f4883575-f600-4012-ac6d-47dfdf2a1394\" (UID: \"f4883575-f600-4012-ac6d-47dfdf2a1394\") " Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.733202 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities" (OuterVolumeSpecName: "utilities") pod "f4883575-f600-4012-ac6d-47dfdf2a1394" (UID: "f4883575-f600-4012-ac6d-47dfdf2a1394"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.743827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj" (OuterVolumeSpecName: "kube-api-access-pdrbj") pod "f4883575-f600-4012-ac6d-47dfdf2a1394" (UID: "f4883575-f600-4012-ac6d-47dfdf2a1394"). InnerVolumeSpecName "kube-api-access-pdrbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.816342 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4883575-f600-4012-ac6d-47dfdf2a1394" (UID: "f4883575-f600-4012-ac6d-47dfdf2a1394"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.834359 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdrbj\" (UniqueName: \"kubernetes.io/projected/f4883575-f600-4012-ac6d-47dfdf2a1394-kube-api-access-pdrbj\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.834413 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:43 crc kubenswrapper[4846]: I0320 11:15:43.834426 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4883575-f600-4012-ac6d-47dfdf2a1394-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.207069 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ms9js" event={"ID":"f4883575-f600-4012-ac6d-47dfdf2a1394","Type":"ContainerDied","Data":"011d5b0e8d7b0770eeb885099ab305668a2e82ada46a7eb972f1fc659114af8c"} Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.208285 4846 scope.go:117] "RemoveContainer" containerID="c4920aaef9fc2489b79d885c1945b573a0cebfed39a9ceb6f659750e966456ab" Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.207265 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ms9js" Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.232619 4846 scope.go:117] "RemoveContainer" containerID="977b19b8b56784c5c72bc1a4dec37ac96742a737204fa238d9dd1701059afe13" Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.253457 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.260624 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ms9js"] Mar 20 11:15:44 crc kubenswrapper[4846]: I0320 11:15:44.267101 4846 scope.go:117] "RemoveContainer" containerID="6e3b3becea99460e3829656e31ccd241d0516848dc5bcd7ee238ac7883745cba" Mar 20 11:15:45 crc kubenswrapper[4846]: I0320 11:15:45.337544 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" path="/var/lib/kubelet/pods/f4883575-f600-4012-ac6d-47dfdf2a1394/volumes" Mar 20 11:15:47 crc kubenswrapper[4846]: I0320 11:15:47.235112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" event={"ID":"bd21bfeb-d2b2-4757-aa36-094c039951f0","Type":"ContainerStarted","Data":"cce7df78b60dfad144a70bbd470bbee09fc811de1d53028f619f7956f3a4409c"} Mar 20 11:15:47 crc kubenswrapper[4846]: I0320 11:15:47.236235 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:15:47 crc kubenswrapper[4846]: I0320 11:15:47.785877 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-ht246" Mar 20 11:15:47 crc kubenswrapper[4846]: I0320 11:15:47.801211 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" podStartSLOduration=4.058676848 podStartE2EDuration="50.801184405s" podCreationTimestamp="2026-03-20 11:14:57 +0000 UTC" firstStartedPulling="2026-03-20 11:14:59.977034224 +0000 UTC m=+1071.343442460" lastFinishedPulling="2026-03-20 11:15:46.719541781 +0000 UTC m=+1118.085950017" observedRunningTime="2026-03-20 11:15:47.259557376 +0000 UTC m=+1118.625965672" watchObservedRunningTime="2026-03-20 11:15:47.801184405 +0000 UTC m=+1119.167592641" Mar 20 11:15:47 crc kubenswrapper[4846]: I0320 11:15:47.870723 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-qg5jf" Mar 20 11:15:48 crc kubenswrapper[4846]: I0320 11:15:48.174435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-hgqt9" Mar 20 11:15:58 crc kubenswrapper[4846]: I0320 11:15:58.393583 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-f45j6" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154169 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566756-wtl7r"] Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154604 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="extract-utilities" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154620 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="extract-utilities" Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154656 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154665 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154680 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="extract-content" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154691 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="extract-content" Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154709 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154718 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154731 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="extract-utilities" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154738 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="extract-utilities" Mar 20 11:16:00 crc kubenswrapper[4846]: E0320 11:16:00.154755 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="extract-content" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154762 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="extract-content" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154965 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4883575-f600-4012-ac6d-47dfdf2a1394" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.154993 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebdd44c3-dd2d-4fbf-aeba-1ddf59a23191" containerName="registry-server" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.155827 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.159814 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.159931 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.160046 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.166457 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566756-wtl7r"] Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.306076 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w2kk\" (UniqueName: \"kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk\") pod \"auto-csr-approver-29566756-wtl7r\" (UID: \"bb5b0771-1014-46aa-8a34-6a9812c36603\") " pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.407655 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w2kk\" (UniqueName: \"kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk\") pod \"auto-csr-approver-29566756-wtl7r\" (UID: \"bb5b0771-1014-46aa-8a34-6a9812c36603\") " pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.430503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w2kk\" (UniqueName: \"kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk\") pod \"auto-csr-approver-29566756-wtl7r\" (UID: \"bb5b0771-1014-46aa-8a34-6a9812c36603\") " pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.478196 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:00 crc kubenswrapper[4846]: I0320 11:16:00.934817 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566756-wtl7r"] Mar 20 11:16:01 crc kubenswrapper[4846]: I0320 11:16:01.358720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" event={"ID":"bb5b0771-1014-46aa-8a34-6a9812c36603","Type":"ContainerStarted","Data":"5917e6153aedc76b2d483135798255a0e8cbe3bb97a78609f3e5f5fa564a1f8b"} Mar 20 11:16:03 crc kubenswrapper[4846]: I0320 11:16:03.382872 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb5b0771-1014-46aa-8a34-6a9812c36603" containerID="717c85bd3707f6822dd601a2fc95931bdcf869f40d934b32bb4fe24157405f45" exitCode=0 Mar 20 11:16:03 crc kubenswrapper[4846]: I0320 11:16:03.382986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" event={"ID":"bb5b0771-1014-46aa-8a34-6a9812c36603","Type":"ContainerDied","Data":"717c85bd3707f6822dd601a2fc95931bdcf869f40d934b32bb4fe24157405f45"} Mar 20 11:16:04 crc kubenswrapper[4846]: I0320 11:16:04.688793 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:04 crc kubenswrapper[4846]: I0320 11:16:04.705387 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w2kk\" (UniqueName: \"kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk\") pod \"bb5b0771-1014-46aa-8a34-6a9812c36603\" (UID: \"bb5b0771-1014-46aa-8a34-6a9812c36603\") " Mar 20 11:16:04 crc kubenswrapper[4846]: I0320 11:16:04.713947 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk" (OuterVolumeSpecName: "kube-api-access-6w2kk") pod "bb5b0771-1014-46aa-8a34-6a9812c36603" (UID: "bb5b0771-1014-46aa-8a34-6a9812c36603"). InnerVolumeSpecName "kube-api-access-6w2kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:16:04 crc kubenswrapper[4846]: I0320 11:16:04.807685 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w2kk\" (UniqueName: \"kubernetes.io/projected/bb5b0771-1014-46aa-8a34-6a9812c36603-kube-api-access-6w2kk\") on node \"crc\" DevicePath \"\"" Mar 20 11:16:05 crc kubenswrapper[4846]: I0320 11:16:05.405698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" event={"ID":"bb5b0771-1014-46aa-8a34-6a9812c36603","Type":"ContainerDied","Data":"5917e6153aedc76b2d483135798255a0e8cbe3bb97a78609f3e5f5fa564a1f8b"} Mar 20 11:16:05 crc kubenswrapper[4846]: I0320 11:16:05.405765 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566756-wtl7r" Mar 20 11:16:05 crc kubenswrapper[4846]: I0320 11:16:05.405782 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5917e6153aedc76b2d483135798255a0e8cbe3bb97a78609f3e5f5fa564a1f8b" Mar 20 11:16:05 crc kubenswrapper[4846]: I0320 11:16:05.807111 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566750-dfpjf"] Mar 20 11:16:05 crc kubenswrapper[4846]: I0320 11:16:05.818663 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566750-dfpjf"] Mar 20 11:16:07 crc kubenswrapper[4846]: I0320 11:16:07.334042 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f" path="/var/lib/kubelet/pods/a4ccfcfe-c1a9-4e18-a811-aebbe93a4c9f/volumes" Mar 20 11:16:13 crc kubenswrapper[4846]: I0320 11:16:13.962998 4846 scope.go:117] "RemoveContainer" containerID="72b4af7ccff029fbd1ac3121efe9a09c2853adaf8f43939366dda430bd481890" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.807189 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:15 crc kubenswrapper[4846]: E0320 11:16:15.808492 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5b0771-1014-46aa-8a34-6a9812c36603" containerName="oc" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.808512 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5b0771-1014-46aa-8a34-6a9812c36603" containerName="oc" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.808685 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5b0771-1014-46aa-8a34-6a9812c36603" containerName="oc" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.809702 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.815361 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.815528 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.815720 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gn7ff" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.815945 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.823112 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.931359 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q2qnv"] Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.932597 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.935992 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.956740 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q2qnv"] Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.998974 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:15 crc kubenswrapper[4846]: I0320 11:16:15.999043 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdv8v\" (UniqueName: \"kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.100505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-config\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.100581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.101279 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8xz9\" (UniqueName: \"kubernetes.io/projected/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-kube-api-access-m8xz9\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.101350 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdv8v\" (UniqueName: \"kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.101497 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.102065 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.121097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdv8v\" (UniqueName: \"kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v\") pod \"dnsmasq-dns-675f4bcbfc-dfn7b\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.132172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.203354 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-config\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.203464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8xz9\" (UniqueName: \"kubernetes.io/projected/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-kube-api-access-m8xz9\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.203556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.204429 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-config\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.204740 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.228177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8xz9\" (UniqueName: \"kubernetes.io/projected/e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4-kube-api-access-m8xz9\") pod \"dnsmasq-dns-78dd6ddcc-q2qnv\" (UID: \"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.251802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:16 crc kubenswrapper[4846]: W0320 11:16:16.447660 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83fe6c80_16f7_4748_99fa_dc93f8cfacf2.slice/crio-6e63abd0fc5cf315e8771c3116a21287eb2f18f9f5f9a1e82252be525fc1e026 WatchSource:0}: Error finding container 6e63abd0fc5cf315e8771c3116a21287eb2f18f9f5f9a1e82252be525fc1e026: Status 404 returned error can't find the container with id 6e63abd0fc5cf315e8771c3116a21287eb2f18f9f5f9a1e82252be525fc1e026 Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.449083 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.512988 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" event={"ID":"83fe6c80-16f7-4748-99fa-dc93f8cfacf2","Type":"ContainerStarted","Data":"6e63abd0fc5cf315e8771c3116a21287eb2f18f9f5f9a1e82252be525fc1e026"} Mar 20 11:16:16 crc kubenswrapper[4846]: I0320 11:16:16.730997 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-q2qnv"] Mar 20 11:16:16 crc kubenswrapper[4846]: W0320 11:16:16.735414 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8dacc6e_4efd_4782_b0dd_f942e7b9bfc4.slice/crio-2aa52b2b78db21f1417b86aa08084f11c5d1e7d77c3ba182e12dd1d31dd2b293 WatchSource:0}: Error finding container 2aa52b2b78db21f1417b86aa08084f11c5d1e7d77c3ba182e12dd1d31dd2b293: Status 404 returned error can't find the container with id 2aa52b2b78db21f1417b86aa08084f11c5d1e7d77c3ba182e12dd1d31dd2b293 Mar 20 11:16:17 crc kubenswrapper[4846]: I0320 11:16:17.527657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" event={"ID":"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4","Type":"ContainerStarted","Data":"2aa52b2b78db21f1417b86aa08084f11c5d1e7d77c3ba182e12dd1d31dd2b293"} Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.451295 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.452405 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdv8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-dfn7b_openstack(83fe6c80-16f7-4748-99fa-dc93f8cfacf2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.453610 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.460992 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.461207 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8xz9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-q2qnv_openstack(e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.462467 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" podUID="e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.848429 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" podUID="e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4" Mar 20 11:16:31 crc kubenswrapper[4846]: E0320 11:16:31.848630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" Mar 20 11:16:39 crc kubenswrapper[4846]: I0320 11:16:39.677459 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:16:39 crc kubenswrapper[4846]: I0320 11:16:39.678089 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:16:43 crc kubenswrapper[4846]: I0320 11:16:43.325763 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:16:44 crc kubenswrapper[4846]: I0320 11:16:44.952231 4846 generic.go:334] "Generic (PLEG): container finished" podID="e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4" containerID="3a6830b4ee4cafef24f708b0568ca305f683f0adda87536de36e749934ddee0c" exitCode=0 Mar 20 11:16:44 crc kubenswrapper[4846]: I0320 11:16:44.952828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" event={"ID":"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4","Type":"ContainerDied","Data":"3a6830b4ee4cafef24f708b0568ca305f683f0adda87536de36e749934ddee0c"} Mar 20 11:16:45 crc kubenswrapper[4846]: I0320 11:16:45.963840 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" event={"ID":"e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4","Type":"ContainerStarted","Data":"b9da45f21fa4381422addd9070e1ef44aaa57275512aa236b67401a02779bd12"} Mar 20 11:16:45 crc kubenswrapper[4846]: I0320 11:16:45.964781 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:45 crc kubenswrapper[4846]: I0320 11:16:45.967512 4846 generic.go:334] "Generic (PLEG): container finished" podID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerID="ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2" exitCode=0 Mar 20 11:16:45 crc kubenswrapper[4846]: I0320 11:16:45.967555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" event={"ID":"83fe6c80-16f7-4748-99fa-dc93f8cfacf2","Type":"ContainerDied","Data":"ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2"} Mar 20 11:16:46 crc kubenswrapper[4846]: I0320 11:16:46.026389 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" podStartSLOduration=3.705848189 podStartE2EDuration="31.026359725s" podCreationTimestamp="2026-03-20 11:16:15 +0000 UTC" firstStartedPulling="2026-03-20 11:16:16.738253801 +0000 UTC m=+1148.104662037" lastFinishedPulling="2026-03-20 11:16:44.058765347 +0000 UTC m=+1175.425173573" observedRunningTime="2026-03-20 11:16:45.992638086 +0000 UTC m=+1177.359046362" watchObservedRunningTime="2026-03-20 11:16:46.026359725 +0000 UTC m=+1177.392767961" Mar 20 11:16:46 crc kubenswrapper[4846]: I0320 11:16:46.980778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" event={"ID":"83fe6c80-16f7-4748-99fa-dc93f8cfacf2","Type":"ContainerStarted","Data":"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d"} Mar 20 11:16:46 crc kubenswrapper[4846]: I0320 11:16:46.982146 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:47 crc kubenswrapper[4846]: I0320 11:16:47.008326 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" podStartSLOduration=-9223372004.846483 podStartE2EDuration="32.008292982s" podCreationTimestamp="2026-03-20 11:16:15 +0000 UTC" firstStartedPulling="2026-03-20 11:16:16.450995315 +0000 UTC m=+1147.817403551" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 11:16:46.99857597 +0000 UTC m=+1178.364984206" watchObservedRunningTime="2026-03-20 11:16:47.008292982 +0000 UTC m=+1178.374701258" Mar 20 11:16:52 crc kubenswrapper[4846]: I0320 11:16:51.133085 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:52 crc kubenswrapper[4846]: I0320 11:16:51.255217 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dd6ddcc-q2qnv" Mar 20 11:16:52 crc kubenswrapper[4846]: I0320 11:16:51.308756 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:52 crc kubenswrapper[4846]: I0320 11:16:52.024589 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="dnsmasq-dns" containerID="cri-o://821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d" gracePeriod=10 Mar 20 11:16:52 crc kubenswrapper[4846]: I0320 11:16:52.960792 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.039244 4846 generic.go:334] "Generic (PLEG): container finished" podID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerID="821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d" exitCode=0 Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.039311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" event={"ID":"83fe6c80-16f7-4748-99fa-dc93f8cfacf2","Type":"ContainerDied","Data":"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d"} Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.039348 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" event={"ID":"83fe6c80-16f7-4748-99fa-dc93f8cfacf2","Type":"ContainerDied","Data":"6e63abd0fc5cf315e8771c3116a21287eb2f18f9f5f9a1e82252be525fc1e026"} Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.039378 4846 scope.go:117] "RemoveContainer" containerID="821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.039524 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-dfn7b" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.057806 4846 scope.go:117] "RemoveContainer" containerID="ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.058386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdv8v\" (UniqueName: \"kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v\") pod \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.058600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config\") pod \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\" (UID: \"83fe6c80-16f7-4748-99fa-dc93f8cfacf2\") " Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.067142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v" (OuterVolumeSpecName: "kube-api-access-zdv8v") pod "83fe6c80-16f7-4748-99fa-dc93f8cfacf2" (UID: "83fe6c80-16f7-4748-99fa-dc93f8cfacf2"). InnerVolumeSpecName "kube-api-access-zdv8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.075426 4846 scope.go:117] "RemoveContainer" containerID="821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d" Mar 20 11:16:53 crc kubenswrapper[4846]: E0320 11:16:53.075814 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d\": container with ID starting with 821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d not found: ID does not exist" containerID="821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.075845 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d"} err="failed to get container status \"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d\": rpc error: code = NotFound desc = could not find container \"821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d\": container with ID starting with 821b61ca16e9b4c81c12d0207ac688586bc29433e9a079aeea05e655b734b00d not found: ID does not exist" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.075876 4846 scope.go:117] "RemoveContainer" containerID="ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2" Mar 20 11:16:53 crc kubenswrapper[4846]: E0320 11:16:53.076378 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2\": container with ID starting with ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2 not found: ID does not exist" containerID="ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.076459 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2"} err="failed to get container status \"ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2\": rpc error: code = NotFound desc = could not find container \"ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2\": container with ID starting with ba93127e0e2254a33ea950ce75faac3eb24b090e3d63650e01f082bcaffdefa2 not found: ID does not exist" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.092464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config" (OuterVolumeSpecName: "config") pod "83fe6c80-16f7-4748-99fa-dc93f8cfacf2" (UID: "83fe6c80-16f7-4748-99fa-dc93f8cfacf2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.160537 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdv8v\" (UniqueName: \"kubernetes.io/projected/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-kube-api-access-zdv8v\") on node \"crc\" DevicePath \"\"" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.160606 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83fe6c80-16f7-4748-99fa-dc93f8cfacf2-config\") on node \"crc\" DevicePath \"\"" Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.379192 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:53 crc kubenswrapper[4846]: I0320 11:16:53.385717 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-dfn7b"] Mar 20 11:16:55 crc kubenswrapper[4846]: I0320 11:16:55.331779 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" path="/var/lib/kubelet/pods/83fe6c80-16f7-4748-99fa-dc93f8cfacf2/volumes" Mar 20 11:17:09 crc kubenswrapper[4846]: I0320 11:17:09.678145 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:17:09 crc kubenswrapper[4846]: I0320 11:17:09.679061 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:17:39 crc kubenswrapper[4846]: I0320 11:17:39.677333 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:17:39 crc kubenswrapper[4846]: I0320 11:17:39.677980 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:17:39 crc kubenswrapper[4846]: I0320 11:17:39.678029 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:17:39 crc kubenswrapper[4846]: I0320 11:17:39.678608 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:17:39 crc kubenswrapper[4846]: I0320 11:17:39.678664 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2" gracePeriod=600 Mar 20 11:17:40 crc kubenswrapper[4846]: I0320 11:17:40.418044 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2" exitCode=0 Mar 20 11:17:40 crc kubenswrapper[4846]: I0320 11:17:40.418127 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2"} Mar 20 11:17:40 crc kubenswrapper[4846]: I0320 11:17:40.418792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3"} Mar 20 11:17:40 crc kubenswrapper[4846]: I0320 11:17:40.418839 4846 scope.go:117] "RemoveContainer" containerID="8c88ee3041f4c628f75deb97e7035249048e2c40a916f39dcb9b4b72e1cdef45" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.153140 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566758-jqdqx"] Mar 20 11:18:00 crc kubenswrapper[4846]: E0320 11:18:00.154191 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="dnsmasq-dns" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.154207 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="dnsmasq-dns" Mar 20 11:18:00 crc kubenswrapper[4846]: E0320 11:18:00.154224 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="init" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.154229 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="init" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.154381 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="83fe6c80-16f7-4748-99fa-dc93f8cfacf2" containerName="dnsmasq-dns" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.155011 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.157755 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.163019 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.163186 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.164323 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566758-jqdqx"] Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.226580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t977\" (UniqueName: \"kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977\") pod \"auto-csr-approver-29566758-jqdqx\" (UID: \"821ed48e-93e8-452a-afb1-762951a28477\") " pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.327791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t977\" (UniqueName: \"kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977\") pod \"auto-csr-approver-29566758-jqdqx\" (UID: \"821ed48e-93e8-452a-afb1-762951a28477\") " pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.348305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t977\" (UniqueName: \"kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977\") pod \"auto-csr-approver-29566758-jqdqx\" (UID: \"821ed48e-93e8-452a-afb1-762951a28477\") " pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.490972 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:00 crc kubenswrapper[4846]: I0320 11:18:00.925092 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566758-jqdqx"] Mar 20 11:18:01 crc kubenswrapper[4846]: I0320 11:18:01.603109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" event={"ID":"821ed48e-93e8-452a-afb1-762951a28477","Type":"ContainerStarted","Data":"8f9088dc60a334195ce97fa0d3bcb203c92d01dc20cdd86a2d19d3f2a77cebb6"} Mar 20 11:18:02 crc kubenswrapper[4846]: I0320 11:18:02.620188 4846 generic.go:334] "Generic (PLEG): container finished" podID="821ed48e-93e8-452a-afb1-762951a28477" containerID="f38279164f1a8e4fa85ec2ae79a24bfbcf41a8443aa34bf8df7f21723192a97e" exitCode=0 Mar 20 11:18:02 crc kubenswrapper[4846]: I0320 11:18:02.620297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" event={"ID":"821ed48e-93e8-452a-afb1-762951a28477","Type":"ContainerDied","Data":"f38279164f1a8e4fa85ec2ae79a24bfbcf41a8443aa34bf8df7f21723192a97e"} Mar 20 11:18:03 crc kubenswrapper[4846]: I0320 11:18:03.982592 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.087954 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t977\" (UniqueName: \"kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977\") pod \"821ed48e-93e8-452a-afb1-762951a28477\" (UID: \"821ed48e-93e8-452a-afb1-762951a28477\") " Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.097182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977" (OuterVolumeSpecName: "kube-api-access-7t977") pod "821ed48e-93e8-452a-afb1-762951a28477" (UID: "821ed48e-93e8-452a-afb1-762951a28477"). InnerVolumeSpecName "kube-api-access-7t977". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.190738 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t977\" (UniqueName: \"kubernetes.io/projected/821ed48e-93e8-452a-afb1-762951a28477-kube-api-access-7t977\") on node \"crc\" DevicePath \"\"" Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.638172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" event={"ID":"821ed48e-93e8-452a-afb1-762951a28477","Type":"ContainerDied","Data":"8f9088dc60a334195ce97fa0d3bcb203c92d01dc20cdd86a2d19d3f2a77cebb6"} Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.638263 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f9088dc60a334195ce97fa0d3bcb203c92d01dc20cdd86a2d19d3f2a77cebb6" Mar 20 11:18:04 crc kubenswrapper[4846]: I0320 11:18:04.638275 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566758-jqdqx" Mar 20 11:18:05 crc kubenswrapper[4846]: I0320 11:18:05.069770 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566752-qrzct"] Mar 20 11:18:05 crc kubenswrapper[4846]: I0320 11:18:05.074475 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566752-qrzct"] Mar 20 11:18:05 crc kubenswrapper[4846]: I0320 11:18:05.334273 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd72fbad-b2b4-44b0-8de2-b6209cb883a5" path="/var/lib/kubelet/pods/cd72fbad-b2b4-44b0-8de2-b6209cb883a5/volumes" Mar 20 11:18:14 crc kubenswrapper[4846]: I0320 11:18:14.080704 4846 scope.go:117] "RemoveContainer" containerID="f97f4aac2dc141a4c63059d1b591089ffec6a0503f2c63998c10ca35e916f80d" Mar 20 11:19:39 crc kubenswrapper[4846]: I0320 11:19:39.678542 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:19:39 crc kubenswrapper[4846]: I0320 11:19:39.679145 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.158602 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566760-vq96x"] Mar 20 11:20:00 crc kubenswrapper[4846]: E0320 11:20:00.159885 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821ed48e-93e8-452a-afb1-762951a28477" containerName="oc" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.159920 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="821ed48e-93e8-452a-afb1-762951a28477" containerName="oc" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.160097 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="821ed48e-93e8-452a-afb1-762951a28477" containerName="oc" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.160814 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.163804 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.166205 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566760-vq96x"] Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.169496 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.169698 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.206925 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8ppd\" (UniqueName: \"kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd\") pod \"auto-csr-approver-29566760-vq96x\" (UID: \"5dcaa656-aefe-4a5d-b389-094e81828884\") " pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.308455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8ppd\" (UniqueName: \"kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd\") pod \"auto-csr-approver-29566760-vq96x\" (UID: \"5dcaa656-aefe-4a5d-b389-094e81828884\") " pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.340819 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8ppd\" (UniqueName: \"kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd\") pod \"auto-csr-approver-29566760-vq96x\" (UID: \"5dcaa656-aefe-4a5d-b389-094e81828884\") " pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.484324 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:00 crc kubenswrapper[4846]: I0320 11:20:00.952025 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566760-vq96x"] Mar 20 11:20:01 crc kubenswrapper[4846]: I0320 11:20:01.533087 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566760-vq96x" event={"ID":"5dcaa656-aefe-4a5d-b389-094e81828884","Type":"ContainerStarted","Data":"c8630b88e663d7d0db1449ca32b754d390f0f12fa692ba5a4a46e76d838d06d5"} Mar 20 11:20:03 crc kubenswrapper[4846]: I0320 11:20:03.546768 4846 generic.go:334] "Generic (PLEG): container finished" podID="5dcaa656-aefe-4a5d-b389-094e81828884" containerID="e204f34cc2842f37fe0886f1eb417ec9fd0a9bd255d0314332585e0e51745cf6" exitCode=0 Mar 20 11:20:03 crc kubenswrapper[4846]: I0320 11:20:03.546827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566760-vq96x" event={"ID":"5dcaa656-aefe-4a5d-b389-094e81828884","Type":"ContainerDied","Data":"e204f34cc2842f37fe0886f1eb417ec9fd0a9bd255d0314332585e0e51745cf6"} Mar 20 11:20:04 crc kubenswrapper[4846]: I0320 11:20:04.845075 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:04 crc kubenswrapper[4846]: I0320 11:20:04.878491 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8ppd\" (UniqueName: \"kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd\") pod \"5dcaa656-aefe-4a5d-b389-094e81828884\" (UID: \"5dcaa656-aefe-4a5d-b389-094e81828884\") " Mar 20 11:20:04 crc kubenswrapper[4846]: I0320 11:20:04.889659 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd" (OuterVolumeSpecName: "kube-api-access-p8ppd") pod "5dcaa656-aefe-4a5d-b389-094e81828884" (UID: "5dcaa656-aefe-4a5d-b389-094e81828884"). InnerVolumeSpecName "kube-api-access-p8ppd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:20:04 crc kubenswrapper[4846]: I0320 11:20:04.979961 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8ppd\" (UniqueName: \"kubernetes.io/projected/5dcaa656-aefe-4a5d-b389-094e81828884-kube-api-access-p8ppd\") on node \"crc\" DevicePath \"\"" Mar 20 11:20:05 crc kubenswrapper[4846]: I0320 11:20:05.563132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566760-vq96x" event={"ID":"5dcaa656-aefe-4a5d-b389-094e81828884","Type":"ContainerDied","Data":"c8630b88e663d7d0db1449ca32b754d390f0f12fa692ba5a4a46e76d838d06d5"} Mar 20 11:20:05 crc kubenswrapper[4846]: I0320 11:20:05.563187 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8630b88e663d7d0db1449ca32b754d390f0f12fa692ba5a4a46e76d838d06d5" Mar 20 11:20:05 crc kubenswrapper[4846]: I0320 11:20:05.563255 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566760-vq96x" Mar 20 11:20:05 crc kubenswrapper[4846]: I0320 11:20:05.933836 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566754-fsgs8"] Mar 20 11:20:05 crc kubenswrapper[4846]: I0320 11:20:05.938551 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566754-fsgs8"] Mar 20 11:20:07 crc kubenswrapper[4846]: I0320 11:20:07.333829 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86e748e6-2de1-464c-ad92-02ad94876039" path="/var/lib/kubelet/pods/86e748e6-2de1-464c-ad92-02ad94876039/volumes" Mar 20 11:20:09 crc kubenswrapper[4846]: I0320 11:20:09.677533 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:20:09 crc kubenswrapper[4846]: I0320 11:20:09.677986 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:20:14 crc kubenswrapper[4846]: I0320 11:20:14.164144 4846 scope.go:117] "RemoveContainer" containerID="71ea7812a6e285a3074f0f4b4ce22be0911fc05882a2d58ee727fd07d8d1beb3" Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.677529 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.678349 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.678411 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.679222 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.679296 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3" gracePeriod=600 Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.992169 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3"} Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.992891 4846 scope.go:117] "RemoveContainer" containerID="e6ddfc286b20537016cabc7ce2b096e102188ab907cfef0f23e1ed07324c73f2" Mar 20 11:20:39 crc kubenswrapper[4846]: I0320 11:20:39.992139 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3" exitCode=0 Mar 20 11:20:41 crc kubenswrapper[4846]: I0320 11:20:41.006318 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2"} Mar 20 11:21:14 crc kubenswrapper[4846]: I0320 11:21:14.225473 4846 scope.go:117] "RemoveContainer" containerID="7a54235385159617dda01febaecefeb85725d80b63557be9161dbaef724180f7" Mar 20 11:21:14 crc kubenswrapper[4846]: I0320 11:21:14.255862 4846 scope.go:117] "RemoveContainer" containerID="8e4b497960aa01e3270d628196087fbfd0a7eb853b9fe0f95f6a0ae8c6ce790e" Mar 20 11:21:14 crc kubenswrapper[4846]: I0320 11:21:14.285153 4846 scope.go:117] "RemoveContainer" containerID="9086ed4556a5802cbd4d6f2f83c978521bc4e3ae1b43755ec53a2ada4b7cdb15" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.069269 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:36 crc kubenswrapper[4846]: E0320 11:21:36.072330 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcaa656-aefe-4a5d-b389-094e81828884" containerName="oc" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.072432 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcaa656-aefe-4a5d-b389-094e81828884" containerName="oc" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.072697 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcaa656-aefe-4a5d-b389-094e81828884" containerName="oc" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.073991 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.102196 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.207814 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.207883 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.208072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wncj\" (UniqueName: \"kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.310303 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wncj\" (UniqueName: \"kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.310392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.310429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.310993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.311087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.338847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wncj\" (UniqueName: \"kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj\") pod \"redhat-operators-qmqcx\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.410802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:36 crc kubenswrapper[4846]: I0320 11:21:36.882769 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:37 crc kubenswrapper[4846]: I0320 11:21:37.477890 4846 generic.go:334] "Generic (PLEG): container finished" podID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerID="f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783" exitCode=0 Mar 20 11:21:37 crc kubenswrapper[4846]: I0320 11:21:37.477967 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerDied","Data":"f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783"} Mar 20 11:21:37 crc kubenswrapper[4846]: I0320 11:21:37.477998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerStarted","Data":"b464ba07d794b746e0de23bb35fe0b0a89e3e1350e7167f0f9929f1ba0bb7b0d"} Mar 20 11:21:39 crc kubenswrapper[4846]: I0320 11:21:39.502031 4846 generic.go:334] "Generic (PLEG): container finished" podID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerID="8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23" exitCode=0 Mar 20 11:21:39 crc kubenswrapper[4846]: I0320 11:21:39.502258 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerDied","Data":"8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23"} Mar 20 11:21:40 crc kubenswrapper[4846]: I0320 11:21:40.523350 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerStarted","Data":"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988"} Mar 20 11:21:40 crc kubenswrapper[4846]: I0320 11:21:40.550973 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qmqcx" podStartSLOduration=2.107350518 podStartE2EDuration="4.550953682s" podCreationTimestamp="2026-03-20 11:21:36 +0000 UTC" firstStartedPulling="2026-03-20 11:21:37.479958631 +0000 UTC m=+1468.846366877" lastFinishedPulling="2026-03-20 11:21:39.923561805 +0000 UTC m=+1471.289970041" observedRunningTime="2026-03-20 11:21:40.543870612 +0000 UTC m=+1471.910278858" watchObservedRunningTime="2026-03-20 11:21:40.550953682 +0000 UTC m=+1471.917361918" Mar 20 11:21:46 crc kubenswrapper[4846]: I0320 11:21:46.411214 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:46 crc kubenswrapper[4846]: I0320 11:21:46.412194 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:47 crc kubenswrapper[4846]: I0320 11:21:47.466251 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qmqcx" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="registry-server" probeResult="failure" output=< Mar 20 11:21:47 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:21:47 crc kubenswrapper[4846]: > Mar 20 11:21:56 crc kubenswrapper[4846]: I0320 11:21:56.463691 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:56 crc kubenswrapper[4846]: I0320 11:21:56.520854 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:56 crc kubenswrapper[4846]: I0320 11:21:56.709576 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:57 crc kubenswrapper[4846]: I0320 11:21:57.661143 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qmqcx" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="registry-server" containerID="cri-o://c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988" gracePeriod=2 Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.112612 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.185339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wncj\" (UniqueName: \"kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj\") pod \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.185493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content\") pod \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.185610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities\") pod \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\" (UID: \"616a4dfe-489b-4d58-9ab8-2648063fbd4a\") " Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.187577 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities" (OuterVolumeSpecName: "utilities") pod "616a4dfe-489b-4d58-9ab8-2648063fbd4a" (UID: "616a4dfe-489b-4d58-9ab8-2648063fbd4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.195622 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj" (OuterVolumeSpecName: "kube-api-access-4wncj") pod "616a4dfe-489b-4d58-9ab8-2648063fbd4a" (UID: "616a4dfe-489b-4d58-9ab8-2648063fbd4a"). InnerVolumeSpecName "kube-api-access-4wncj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.287418 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.287515 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wncj\" (UniqueName: \"kubernetes.io/projected/616a4dfe-489b-4d58-9ab8-2648063fbd4a-kube-api-access-4wncj\") on node \"crc\" DevicePath \"\"" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.334310 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "616a4dfe-489b-4d58-9ab8-2648063fbd4a" (UID: "616a4dfe-489b-4d58-9ab8-2648063fbd4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.389095 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616a4dfe-489b-4d58-9ab8-2648063fbd4a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.673511 4846 generic.go:334] "Generic (PLEG): container finished" podID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerID="c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988" exitCode=0 Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.673577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerDied","Data":"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988"} Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.673601 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qmqcx" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.673627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qmqcx" event={"ID":"616a4dfe-489b-4d58-9ab8-2648063fbd4a","Type":"ContainerDied","Data":"b464ba07d794b746e0de23bb35fe0b0a89e3e1350e7167f0f9929f1ba0bb7b0d"} Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.673675 4846 scope.go:117] "RemoveContainer" containerID="c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.722597 4846 scope.go:117] "RemoveContainer" containerID="8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.723686 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.730752 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qmqcx"] Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.746371 4846 scope.go:117] "RemoveContainer" containerID="f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.769305 4846 scope.go:117] "RemoveContainer" containerID="c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988" Mar 20 11:21:58 crc kubenswrapper[4846]: E0320 11:21:58.769831 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988\": container with ID starting with c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988 not found: ID does not exist" containerID="c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.769909 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988"} err="failed to get container status \"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988\": rpc error: code = NotFound desc = could not find container \"c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988\": container with ID starting with c048afde2f6e223cd76f2b82f5224c3413185ec35dc5eddf9da50e3ce0019988 not found: ID does not exist" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.769956 4846 scope.go:117] "RemoveContainer" containerID="8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23" Mar 20 11:21:58 crc kubenswrapper[4846]: E0320 11:21:58.770363 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23\": container with ID starting with 8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23 not found: ID does not exist" containerID="8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.770403 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23"} err="failed to get container status \"8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23\": rpc error: code = NotFound desc = could not find container \"8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23\": container with ID starting with 8659f1d5829cf6a7f3097f51be73c15b95d7f7f222bfdccc7de3e2d779bbdd23 not found: ID does not exist" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.770433 4846 scope.go:117] "RemoveContainer" containerID="f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783" Mar 20 11:21:58 crc kubenswrapper[4846]: E0320 11:21:58.770704 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783\": container with ID starting with f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783 not found: ID does not exist" containerID="f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783" Mar 20 11:21:58 crc kubenswrapper[4846]: I0320 11:21:58.770733 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783"} err="failed to get container status \"f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783\": rpc error: code = NotFound desc = could not find container \"f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783\": container with ID starting with f37ddfb9f1fa76924d3e26ffec5e1aa8e861cec3fef9a796dc484401c472d783 not found: ID does not exist" Mar 20 11:21:59 crc kubenswrapper[4846]: I0320 11:21:59.339148 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" path="/var/lib/kubelet/pods/616a4dfe-489b-4d58-9ab8-2648063fbd4a/volumes" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.152390 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566762-8l7rd"] Mar 20 11:22:00 crc kubenswrapper[4846]: E0320 11:22:00.153127 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="extract-utilities" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.153154 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="extract-utilities" Mar 20 11:22:00 crc kubenswrapper[4846]: E0320 11:22:00.153179 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="extract-content" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.153190 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="extract-content" Mar 20 11:22:00 crc kubenswrapper[4846]: E0320 11:22:00.153218 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="registry-server" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.153231 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="registry-server" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.153462 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="616a4dfe-489b-4d58-9ab8-2648063fbd4a" containerName="registry-server" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.154272 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.157728 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.157772 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.158725 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.203786 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566762-8l7rd"] Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.226049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xldl\" (UniqueName: \"kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl\") pod \"auto-csr-approver-29566762-8l7rd\" (UID: \"8f490d18-0061-48ea-aa5c-dc0e5118e028\") " pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.328012 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xldl\" (UniqueName: \"kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl\") pod \"auto-csr-approver-29566762-8l7rd\" (UID: \"8f490d18-0061-48ea-aa5c-dc0e5118e028\") " pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.348650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xldl\" (UniqueName: \"kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl\") pod \"auto-csr-approver-29566762-8l7rd\" (UID: \"8f490d18-0061-48ea-aa5c-dc0e5118e028\") " pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.478115 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.921253 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566762-8l7rd"] Mar 20 11:22:00 crc kubenswrapper[4846]: I0320 11:22:00.930778 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:22:01 crc kubenswrapper[4846]: I0320 11:22:01.718680 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" event={"ID":"8f490d18-0061-48ea-aa5c-dc0e5118e028","Type":"ContainerStarted","Data":"cbc52a8062869b721d0dc564c3415a90bc31846ce0b857eadf345e31d085730d"} Mar 20 11:22:03 crc kubenswrapper[4846]: I0320 11:22:03.742058 4846 generic.go:334] "Generic (PLEG): container finished" podID="8f490d18-0061-48ea-aa5c-dc0e5118e028" containerID="a3be040ae25d0267a70d7ea393eb534cd49b73733c6b684ce5f3de4e8f674880" exitCode=0 Mar 20 11:22:03 crc kubenswrapper[4846]: I0320 11:22:03.742206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" event={"ID":"8f490d18-0061-48ea-aa5c-dc0e5118e028","Type":"ContainerDied","Data":"a3be040ae25d0267a70d7ea393eb534cd49b73733c6b684ce5f3de4e8f674880"} Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.064468 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.213620 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xldl\" (UniqueName: \"kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl\") pod \"8f490d18-0061-48ea-aa5c-dc0e5118e028\" (UID: \"8f490d18-0061-48ea-aa5c-dc0e5118e028\") " Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.220005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl" (OuterVolumeSpecName: "kube-api-access-2xldl") pod "8f490d18-0061-48ea-aa5c-dc0e5118e028" (UID: "8f490d18-0061-48ea-aa5c-dc0e5118e028"). InnerVolumeSpecName "kube-api-access-2xldl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.316662 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xldl\" (UniqueName: \"kubernetes.io/projected/8f490d18-0061-48ea-aa5c-dc0e5118e028-kube-api-access-2xldl\") on node \"crc\" DevicePath \"\"" Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.759593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" event={"ID":"8f490d18-0061-48ea-aa5c-dc0e5118e028","Type":"ContainerDied","Data":"cbc52a8062869b721d0dc564c3415a90bc31846ce0b857eadf345e31d085730d"} Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.759650 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc52a8062869b721d0dc564c3415a90bc31846ce0b857eadf345e31d085730d" Mar 20 11:22:05 crc kubenswrapper[4846]: I0320 11:22:05.759672 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566762-8l7rd" Mar 20 11:22:06 crc kubenswrapper[4846]: I0320 11:22:06.136243 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566756-wtl7r"] Mar 20 11:22:06 crc kubenswrapper[4846]: I0320 11:22:06.141085 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566756-wtl7r"] Mar 20 11:22:07 crc kubenswrapper[4846]: I0320 11:22:07.339776 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5b0771-1014-46aa-8a34-6a9812c36603" path="/var/lib/kubelet/pods/bb5b0771-1014-46aa-8a34-6a9812c36603/volumes" Mar 20 11:22:14 crc kubenswrapper[4846]: I0320 11:22:14.361735 4846 scope.go:117] "RemoveContainer" containerID="717c85bd3707f6822dd601a2fc95931bdcf869f40d934b32bb4fe24157405f45" Mar 20 11:23:09 crc kubenswrapper[4846]: I0320 11:23:09.678193 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:23:09 crc kubenswrapper[4846]: I0320 11:23:09.679123 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:23:39 crc kubenswrapper[4846]: I0320 11:23:39.678261 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:23:39 crc kubenswrapper[4846]: I0320 11:23:39.680322 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.146077 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566764-5ngks"] Mar 20 11:24:00 crc kubenswrapper[4846]: E0320 11:24:00.147445 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f490d18-0061-48ea-aa5c-dc0e5118e028" containerName="oc" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.147472 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f490d18-0061-48ea-aa5c-dc0e5118e028" containerName="oc" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.147691 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f490d18-0061-48ea-aa5c-dc0e5118e028" containerName="oc" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.148434 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.152031 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.154990 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.154990 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.157505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566764-5ngks"] Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.264727 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtglv\" (UniqueName: \"kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv\") pod \"auto-csr-approver-29566764-5ngks\" (UID: \"0b73c17a-882b-4565-bc01-ddc1f6576c0e\") " pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.366808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtglv\" (UniqueName: \"kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv\") pod \"auto-csr-approver-29566764-5ngks\" (UID: \"0b73c17a-882b-4565-bc01-ddc1f6576c0e\") " pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.387002 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtglv\" (UniqueName: \"kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv\") pod \"auto-csr-approver-29566764-5ngks\" (UID: \"0b73c17a-882b-4565-bc01-ddc1f6576c0e\") " pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.467861 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:00 crc kubenswrapper[4846]: I0320 11:24:00.990417 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566764-5ngks"] Mar 20 11:24:01 crc kubenswrapper[4846]: I0320 11:24:01.997282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566764-5ngks" event={"ID":"0b73c17a-882b-4565-bc01-ddc1f6576c0e","Type":"ContainerStarted","Data":"be96357bafa73f6b59c6a2d11602a517866b8fd2bf4bc75bd3054d800359012a"} Mar 20 11:24:03 crc kubenswrapper[4846]: I0320 11:24:03.010684 4846 generic.go:334] "Generic (PLEG): container finished" podID="0b73c17a-882b-4565-bc01-ddc1f6576c0e" containerID="cbeec4103f0bc3a35efe6d52b80ba24cc8271da7cf1e9d5f7c630e26e97d6abc" exitCode=0 Mar 20 11:24:03 crc kubenswrapper[4846]: I0320 11:24:03.010834 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566764-5ngks" event={"ID":"0b73c17a-882b-4565-bc01-ddc1f6576c0e","Type":"ContainerDied","Data":"cbeec4103f0bc3a35efe6d52b80ba24cc8271da7cf1e9d5f7c630e26e97d6abc"} Mar 20 11:24:04 crc kubenswrapper[4846]: I0320 11:24:04.358554 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:04 crc kubenswrapper[4846]: I0320 11:24:04.536480 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtglv\" (UniqueName: \"kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv\") pod \"0b73c17a-882b-4565-bc01-ddc1f6576c0e\" (UID: \"0b73c17a-882b-4565-bc01-ddc1f6576c0e\") " Mar 20 11:24:04 crc kubenswrapper[4846]: I0320 11:24:04.544318 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv" (OuterVolumeSpecName: "kube-api-access-gtglv") pod "0b73c17a-882b-4565-bc01-ddc1f6576c0e" (UID: "0b73c17a-882b-4565-bc01-ddc1f6576c0e"). InnerVolumeSpecName "kube-api-access-gtglv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:24:04 crc kubenswrapper[4846]: I0320 11:24:04.638667 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtglv\" (UniqueName: \"kubernetes.io/projected/0b73c17a-882b-4565-bc01-ddc1f6576c0e-kube-api-access-gtglv\") on node \"crc\" DevicePath \"\"" Mar 20 11:24:05 crc kubenswrapper[4846]: I0320 11:24:05.028940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566764-5ngks" event={"ID":"0b73c17a-882b-4565-bc01-ddc1f6576c0e","Type":"ContainerDied","Data":"be96357bafa73f6b59c6a2d11602a517866b8fd2bf4bc75bd3054d800359012a"} Mar 20 11:24:05 crc kubenswrapper[4846]: I0320 11:24:05.029366 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be96357bafa73f6b59c6a2d11602a517866b8fd2bf4bc75bd3054d800359012a" Mar 20 11:24:05 crc kubenswrapper[4846]: I0320 11:24:05.029434 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566764-5ngks" Mar 20 11:24:05 crc kubenswrapper[4846]: I0320 11:24:05.436469 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566758-jqdqx"] Mar 20 11:24:05 crc kubenswrapper[4846]: I0320 11:24:05.442340 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566758-jqdqx"] Mar 20 11:24:07 crc kubenswrapper[4846]: I0320 11:24:07.331283 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="821ed48e-93e8-452a-afb1-762951a28477" path="/var/lib/kubelet/pods/821ed48e-93e8-452a-afb1-762951a28477/volumes" Mar 20 11:24:09 crc kubenswrapper[4846]: I0320 11:24:09.677912 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:24:09 crc kubenswrapper[4846]: I0320 11:24:09.677987 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:24:09 crc kubenswrapper[4846]: I0320 11:24:09.678047 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:24:09 crc kubenswrapper[4846]: I0320 11:24:09.678867 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:24:09 crc kubenswrapper[4846]: I0320 11:24:09.678937 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" gracePeriod=600 Mar 20 11:24:09 crc kubenswrapper[4846]: E0320 11:24:09.820339 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:24:10 crc kubenswrapper[4846]: I0320 11:24:10.074490 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" exitCode=0 Mar 20 11:24:10 crc kubenswrapper[4846]: I0320 11:24:10.074566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2"} Mar 20 11:24:10 crc kubenswrapper[4846]: I0320 11:24:10.074684 4846 scope.go:117] "RemoveContainer" containerID="1bde91157e0d2e23c90e4c0f0f65c4f95be63f8d60331c7cab79d607cc492be3" Mar 20 11:24:10 crc kubenswrapper[4846]: I0320 11:24:10.075750 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:24:10 crc kubenswrapper[4846]: E0320 11:24:10.076271 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:24:14 crc kubenswrapper[4846]: I0320 11:24:14.479078 4846 scope.go:117] "RemoveContainer" containerID="f38279164f1a8e4fa85ec2ae79a24bfbcf41a8443aa34bf8df7f21723192a97e" Mar 20 11:24:22 crc kubenswrapper[4846]: I0320 11:24:22.323549 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:24:22 crc kubenswrapper[4846]: E0320 11:24:22.324611 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:24:37 crc kubenswrapper[4846]: I0320 11:24:37.323155 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:24:37 crc kubenswrapper[4846]: E0320 11:24:37.324179 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:24:52 crc kubenswrapper[4846]: I0320 11:24:52.322632 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:24:52 crc kubenswrapper[4846]: E0320 11:24:52.324540 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:25:06 crc kubenswrapper[4846]: I0320 11:25:06.324043 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:25:06 crc kubenswrapper[4846]: E0320 11:25:06.327086 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:25:17 crc kubenswrapper[4846]: I0320 11:25:17.323570 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:25:17 crc kubenswrapper[4846]: E0320 11:25:17.324609 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:25:28 crc kubenswrapper[4846]: I0320 11:25:28.323336 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:25:28 crc kubenswrapper[4846]: E0320 11:25:28.324393 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:25:42 crc kubenswrapper[4846]: I0320 11:25:42.323470 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:25:42 crc kubenswrapper[4846]: E0320 11:25:42.325103 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:25:56 crc kubenswrapper[4846]: I0320 11:25:56.323696 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:25:56 crc kubenswrapper[4846]: E0320 11:25:56.325174 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.146443 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566766-jsqld"] Mar 20 11:26:00 crc kubenswrapper[4846]: E0320 11:26:00.147476 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b73c17a-882b-4565-bc01-ddc1f6576c0e" containerName="oc" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.147494 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b73c17a-882b-4565-bc01-ddc1f6576c0e" containerName="oc" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.147667 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b73c17a-882b-4565-bc01-ddc1f6576c0e" containerName="oc" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.148317 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.151513 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.151812 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.152005 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.160707 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566766-jsqld"] Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.347694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4wh\" (UniqueName: \"kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh\") pod \"auto-csr-approver-29566766-jsqld\" (UID: \"9dd79e98-23c2-451f-91a6-8bc35bd6380f\") " pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.449969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4wh\" (UniqueName: \"kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh\") pod \"auto-csr-approver-29566766-jsqld\" (UID: \"9dd79e98-23c2-451f-91a6-8bc35bd6380f\") " pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.468877 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4wh\" (UniqueName: \"kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh\") pod \"auto-csr-approver-29566766-jsqld\" (UID: \"9dd79e98-23c2-451f-91a6-8bc35bd6380f\") " pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.469330 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.914174 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566766-jsqld"] Mar 20 11:26:00 crc kubenswrapper[4846]: W0320 11:26:00.919953 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dd79e98_23c2_451f_91a6_8bc35bd6380f.slice/crio-149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8 WatchSource:0}: Error finding container 149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8: Status 404 returned error can't find the container with id 149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8 Mar 20 11:26:00 crc kubenswrapper[4846]: I0320 11:26:00.983104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566766-jsqld" event={"ID":"9dd79e98-23c2-451f-91a6-8bc35bd6380f","Type":"ContainerStarted","Data":"149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8"} Mar 20 11:26:03 crc kubenswrapper[4846]: I0320 11:26:03.002044 4846 generic.go:334] "Generic (PLEG): container finished" podID="9dd79e98-23c2-451f-91a6-8bc35bd6380f" containerID="cb5e084d7f0ee267eb11ba352ed35f56d4ba6ff7c7dfb1951d987e2d534a02aa" exitCode=0 Mar 20 11:26:03 crc kubenswrapper[4846]: I0320 11:26:03.002108 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566766-jsqld" event={"ID":"9dd79e98-23c2-451f-91a6-8bc35bd6380f","Type":"ContainerDied","Data":"cb5e084d7f0ee267eb11ba352ed35f56d4ba6ff7c7dfb1951d987e2d534a02aa"} Mar 20 11:26:04 crc kubenswrapper[4846]: I0320 11:26:04.302846 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:04 crc kubenswrapper[4846]: I0320 11:26:04.415190 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp4wh\" (UniqueName: \"kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh\") pod \"9dd79e98-23c2-451f-91a6-8bc35bd6380f\" (UID: \"9dd79e98-23c2-451f-91a6-8bc35bd6380f\") " Mar 20 11:26:04 crc kubenswrapper[4846]: I0320 11:26:04.422681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh" (OuterVolumeSpecName: "kube-api-access-wp4wh") pod "9dd79e98-23c2-451f-91a6-8bc35bd6380f" (UID: "9dd79e98-23c2-451f-91a6-8bc35bd6380f"). InnerVolumeSpecName "kube-api-access-wp4wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:26:04 crc kubenswrapper[4846]: I0320 11:26:04.516883 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp4wh\" (UniqueName: \"kubernetes.io/projected/9dd79e98-23c2-451f-91a6-8bc35bd6380f-kube-api-access-wp4wh\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:05 crc kubenswrapper[4846]: I0320 11:26:05.021369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566766-jsqld" event={"ID":"9dd79e98-23c2-451f-91a6-8bc35bd6380f","Type":"ContainerDied","Data":"149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8"} Mar 20 11:26:05 crc kubenswrapper[4846]: I0320 11:26:05.021427 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="149af4fc4f8296c8b03ec2b4fcd5b50e58519de5c9ebdf96379dd36b5bf7bfb8" Mar 20 11:26:05 crc kubenswrapper[4846]: I0320 11:26:05.021437 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566766-jsqld" Mar 20 11:26:05 crc kubenswrapper[4846]: I0320 11:26:05.377709 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566760-vq96x"] Mar 20 11:26:05 crc kubenswrapper[4846]: I0320 11:26:05.382954 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566760-vq96x"] Mar 20 11:26:07 crc kubenswrapper[4846]: I0320 11:26:07.333356 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dcaa656-aefe-4a5d-b389-094e81828884" path="/var/lib/kubelet/pods/5dcaa656-aefe-4a5d-b389-094e81828884/volumes" Mar 20 11:26:11 crc kubenswrapper[4846]: I0320 11:26:11.323988 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:26:11 crc kubenswrapper[4846]: E0320 11:26:11.324874 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:26:14 crc kubenswrapper[4846]: I0320 11:26:14.585672 4846 scope.go:117] "RemoveContainer" containerID="e204f34cc2842f37fe0886f1eb417ec9fd0a9bd255d0314332585e0e51745cf6" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.462113 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:16 crc kubenswrapper[4846]: E0320 11:26:16.463085 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd79e98-23c2-451f-91a6-8bc35bd6380f" containerName="oc" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.463103 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd79e98-23c2-451f-91a6-8bc35bd6380f" containerName="oc" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.463299 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd79e98-23c2-451f-91a6-8bc35bd6380f" containerName="oc" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.464569 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.472169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.518010 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.518247 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.518295 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tstl\" (UniqueName: \"kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.619455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.619514 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tstl\" (UniqueName: \"kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.619535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.620147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.620387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.646838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tstl\" (UniqueName: \"kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl\") pod \"certified-operators-5lhcz\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:16 crc kubenswrapper[4846]: I0320 11:26:16.790690 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:17 crc kubenswrapper[4846]: I0320 11:26:17.254608 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:18 crc kubenswrapper[4846]: I0320 11:26:18.136474 4846 generic.go:334] "Generic (PLEG): container finished" podID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerID="6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b" exitCode=0 Mar 20 11:26:18 crc kubenswrapper[4846]: I0320 11:26:18.136573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerDied","Data":"6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b"} Mar 20 11:26:18 crc kubenswrapper[4846]: I0320 11:26:18.140215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerStarted","Data":"da56281ea1d92d42f03f267b0b0ad3a384d1e9e07e6c2c9bdc98484eda5135dd"} Mar 20 11:26:19 crc kubenswrapper[4846]: I0320 11:26:19.150419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerStarted","Data":"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad"} Mar 20 11:26:20 crc kubenswrapper[4846]: I0320 11:26:20.160715 4846 generic.go:334] "Generic (PLEG): container finished" podID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerID="65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad" exitCode=0 Mar 20 11:26:20 crc kubenswrapper[4846]: I0320 11:26:20.160861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerDied","Data":"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad"} Mar 20 11:26:22 crc kubenswrapper[4846]: I0320 11:26:22.181823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerStarted","Data":"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef"} Mar 20 11:26:22 crc kubenswrapper[4846]: I0320 11:26:22.210129 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5lhcz" podStartSLOduration=3.329420935 podStartE2EDuration="6.210106694s" podCreationTimestamp="2026-03-20 11:26:16 +0000 UTC" firstStartedPulling="2026-03-20 11:26:18.139033046 +0000 UTC m=+1749.505441282" lastFinishedPulling="2026-03-20 11:26:21.019718805 +0000 UTC m=+1752.386127041" observedRunningTime="2026-03-20 11:26:22.20338608 +0000 UTC m=+1753.569794316" watchObservedRunningTime="2026-03-20 11:26:22.210106694 +0000 UTC m=+1753.576514930" Mar 20 11:26:25 crc kubenswrapper[4846]: I0320 11:26:25.324110 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:26:25 crc kubenswrapper[4846]: E0320 11:26:25.324891 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:26:26 crc kubenswrapper[4846]: I0320 11:26:26.791269 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:26 crc kubenswrapper[4846]: I0320 11:26:26.791328 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:26 crc kubenswrapper[4846]: I0320 11:26:26.839057 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:27 crc kubenswrapper[4846]: I0320 11:26:27.278587 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:27 crc kubenswrapper[4846]: I0320 11:26:27.346973 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.245154 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5lhcz" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="registry-server" containerID="cri-o://a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef" gracePeriod=2 Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.711013 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.843821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities\") pod \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.843923 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tstl\" (UniqueName: \"kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl\") pod \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.844001 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content\") pod \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\" (UID: \"e0fd6ebf-1891-4834-854f-52bf0b7969bf\") " Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.845128 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities" (OuterVolumeSpecName: "utilities") pod "e0fd6ebf-1891-4834-854f-52bf0b7969bf" (UID: "e0fd6ebf-1891-4834-854f-52bf0b7969bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.848297 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.851498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl" (OuterVolumeSpecName: "kube-api-access-2tstl") pod "e0fd6ebf-1891-4834-854f-52bf0b7969bf" (UID: "e0fd6ebf-1891-4834-854f-52bf0b7969bf"). InnerVolumeSpecName "kube-api-access-2tstl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:26:29 crc kubenswrapper[4846]: I0320 11:26:29.949102 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tstl\" (UniqueName: \"kubernetes.io/projected/e0fd6ebf-1891-4834-854f-52bf0b7969bf-kube-api-access-2tstl\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.177212 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0fd6ebf-1891-4834-854f-52bf0b7969bf" (UID: "e0fd6ebf-1891-4834-854f-52bf0b7969bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.265096 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fd6ebf-1891-4834-854f-52bf0b7969bf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.265766 4846 generic.go:334] "Generic (PLEG): container finished" podID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerID="a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef" exitCode=0 Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.265857 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerDied","Data":"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef"} Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.265918 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lhcz" event={"ID":"e0fd6ebf-1891-4834-854f-52bf0b7969bf","Type":"ContainerDied","Data":"da56281ea1d92d42f03f267b0b0ad3a384d1e9e07e6c2c9bdc98484eda5135dd"} Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.265945 4846 scope.go:117] "RemoveContainer" containerID="a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.266816 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lhcz" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.290169 4846 scope.go:117] "RemoveContainer" containerID="65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.314112 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.317403 4846 scope.go:117] "RemoveContainer" containerID="6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.325615 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5lhcz"] Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.351735 4846 scope.go:117] "RemoveContainer" containerID="a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef" Mar 20 11:26:30 crc kubenswrapper[4846]: E0320 11:26:30.352138 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef\": container with ID starting with a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef not found: ID does not exist" containerID="a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.352189 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef"} err="failed to get container status \"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef\": rpc error: code = NotFound desc = could not find container \"a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef\": container with ID starting with a93f0de51032ce5375fd20d04c04b4b5fe2b8082608a210b6487e5efaa71d8ef not found: ID does not exist" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.352263 4846 scope.go:117] "RemoveContainer" containerID="65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad" Mar 20 11:26:30 crc kubenswrapper[4846]: E0320 11:26:30.352538 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad\": container with ID starting with 65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad not found: ID does not exist" containerID="65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.352599 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad"} err="failed to get container status \"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad\": rpc error: code = NotFound desc = could not find container \"65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad\": container with ID starting with 65b5a91570af515668f8f2d85c61fadcc977bef2fc36f094b80d4d4e98d41bad not found: ID does not exist" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.352619 4846 scope.go:117] "RemoveContainer" containerID="6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b" Mar 20 11:26:30 crc kubenswrapper[4846]: E0320 11:26:30.352865 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b\": container with ID starting with 6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b not found: ID does not exist" containerID="6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b" Mar 20 11:26:30 crc kubenswrapper[4846]: I0320 11:26:30.352927 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b"} err="failed to get container status \"6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b\": rpc error: code = NotFound desc = could not find container \"6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b\": container with ID starting with 6bdd0bc6212adf5913a87f3c0b49ce59dbda05701ed2b0f0c7d8bf0f84826b3b not found: ID does not exist" Mar 20 11:26:31 crc kubenswrapper[4846]: I0320 11:26:31.333082 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" path="/var/lib/kubelet/pods/e0fd6ebf-1891-4834-854f-52bf0b7969bf/volumes" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.325519 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:26:37 crc kubenswrapper[4846]: E0320 11:26:37.326723 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.381059 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7nrmp"] Mar 20 11:26:37 crc kubenswrapper[4846]: E0320 11:26:37.381558 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="extract-utilities" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.381591 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="extract-utilities" Mar 20 11:26:37 crc kubenswrapper[4846]: E0320 11:26:37.381616 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="extract-content" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.381627 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="extract-content" Mar 20 11:26:37 crc kubenswrapper[4846]: E0320 11:26:37.381667 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="registry-server" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.381679 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="registry-server" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.381980 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0fd6ebf-1891-4834-854f-52bf0b7969bf" containerName="registry-server" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.387270 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.394395 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7nrmp"] Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.399054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-catalog-content\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.399109 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h59k5\" (UniqueName: \"kubernetes.io/projected/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-kube-api-access-h59k5\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.399349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-utilities\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.501269 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-utilities\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.501370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-catalog-content\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.501414 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h59k5\" (UniqueName: \"kubernetes.io/projected/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-kube-api-access-h59k5\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.501942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-utilities\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.501999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-catalog-content\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.524616 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h59k5\" (UniqueName: \"kubernetes.io/projected/5e831ea5-2e06-4bf6-a2d3-e3fb117168d0-kube-api-access-h59k5\") pod \"community-operators-7nrmp\" (UID: \"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0\") " pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:37 crc kubenswrapper[4846]: I0320 11:26:37.713793 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:38 crc kubenswrapper[4846]: I0320 11:26:38.232964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7nrmp"] Mar 20 11:26:38 crc kubenswrapper[4846]: I0320 11:26:38.338842 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nrmp" event={"ID":"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0","Type":"ContainerStarted","Data":"95c1fd900a0b69e50c54b2eaca0cb31adcfa0b13d2d2898bb985073d5918990c"} Mar 20 11:26:39 crc kubenswrapper[4846]: I0320 11:26:39.363810 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e831ea5-2e06-4bf6-a2d3-e3fb117168d0" containerID="6a756a28dd72361e3c29c969fb5c453d76192a010da0a125ab0e97d3e1ba5173" exitCode=0 Mar 20 11:26:39 crc kubenswrapper[4846]: I0320 11:26:39.364070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nrmp" event={"ID":"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0","Type":"ContainerDied","Data":"6a756a28dd72361e3c29c969fb5c453d76192a010da0a125ab0e97d3e1ba5173"} Mar 20 11:26:43 crc kubenswrapper[4846]: I0320 11:26:43.411600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nrmp" event={"ID":"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0","Type":"ContainerStarted","Data":"b549e4ee28a6e7d8a9f273befd1c006f8f71788019c676829a653bfe7f3b4417"} Mar 20 11:26:44 crc kubenswrapper[4846]: I0320 11:26:44.425771 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e831ea5-2e06-4bf6-a2d3-e3fb117168d0" containerID="b549e4ee28a6e7d8a9f273befd1c006f8f71788019c676829a653bfe7f3b4417" exitCode=0 Mar 20 11:26:44 crc kubenswrapper[4846]: I0320 11:26:44.425844 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nrmp" event={"ID":"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0","Type":"ContainerDied","Data":"b549e4ee28a6e7d8a9f273befd1c006f8f71788019c676829a653bfe7f3b4417"} Mar 20 11:26:45 crc kubenswrapper[4846]: I0320 11:26:45.436272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nrmp" event={"ID":"5e831ea5-2e06-4bf6-a2d3-e3fb117168d0","Type":"ContainerStarted","Data":"12fb93eeb04d5734ed4b9de6e807bbf92a76772676f6920cca2afd395b01ae6e"} Mar 20 11:26:45 crc kubenswrapper[4846]: I0320 11:26:45.461687 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7nrmp" podStartSLOduration=2.834778127 podStartE2EDuration="8.461657361s" podCreationTimestamp="2026-03-20 11:26:37 +0000 UTC" firstStartedPulling="2026-03-20 11:26:39.367098724 +0000 UTC m=+1770.733506960" lastFinishedPulling="2026-03-20 11:26:44.993977958 +0000 UTC m=+1776.360386194" observedRunningTime="2026-03-20 11:26:45.456299661 +0000 UTC m=+1776.822707897" watchObservedRunningTime="2026-03-20 11:26:45.461657361 +0000 UTC m=+1776.828065597" Mar 20 11:26:47 crc kubenswrapper[4846]: I0320 11:26:47.714258 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:47 crc kubenswrapper[4846]: I0320 11:26:47.716012 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:47 crc kubenswrapper[4846]: I0320 11:26:47.789022 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:49 crc kubenswrapper[4846]: I0320 11:26:49.323181 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:26:49 crc kubenswrapper[4846]: E0320 11:26:49.323655 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:26:57 crc kubenswrapper[4846]: I0320 11:26:57.780231 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7nrmp" Mar 20 11:26:57 crc kubenswrapper[4846]: I0320 11:26:57.872631 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7nrmp"] Mar 20 11:26:57 crc kubenswrapper[4846]: I0320 11:26:57.948312 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:26:57 crc kubenswrapper[4846]: I0320 11:26:57.948599 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zb6fs" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="registry-server" containerID="cri-o://0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5" gracePeriod=2 Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.390768 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.506973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities\") pod \"6105d963-9e35-4d0b-897f-8b4488d105ba\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.507125 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content\") pod \"6105d963-9e35-4d0b-897f-8b4488d105ba\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.507177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4j7x\" (UniqueName: \"kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x\") pod \"6105d963-9e35-4d0b-897f-8b4488d105ba\" (UID: \"6105d963-9e35-4d0b-897f-8b4488d105ba\") " Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.507994 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities" (OuterVolumeSpecName: "utilities") pod "6105d963-9e35-4d0b-897f-8b4488d105ba" (UID: "6105d963-9e35-4d0b-897f-8b4488d105ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.516322 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x" (OuterVolumeSpecName: "kube-api-access-p4j7x") pod "6105d963-9e35-4d0b-897f-8b4488d105ba" (UID: "6105d963-9e35-4d0b-897f-8b4488d105ba"). InnerVolumeSpecName "kube-api-access-p4j7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.550791 4846 generic.go:334] "Generic (PLEG): container finished" podID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerID="0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5" exitCode=0 Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.550875 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb6fs" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.550868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerDied","Data":"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5"} Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.550946 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb6fs" event={"ID":"6105d963-9e35-4d0b-897f-8b4488d105ba","Type":"ContainerDied","Data":"a105c0b337bff13800c441a9c0442dbfef2d7470898656e43a5c71cb04108b6a"} Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.550971 4846 scope.go:117] "RemoveContainer" containerID="0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.560680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6105d963-9e35-4d0b-897f-8b4488d105ba" (UID: "6105d963-9e35-4d0b-897f-8b4488d105ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.575447 4846 scope.go:117] "RemoveContainer" containerID="4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.603007 4846 scope.go:117] "RemoveContainer" containerID="4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.612078 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.612131 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4j7x\" (UniqueName: \"kubernetes.io/projected/6105d963-9e35-4d0b-897f-8b4488d105ba-kube-api-access-p4j7x\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.612146 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6105d963-9e35-4d0b-897f-8b4488d105ba-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.629245 4846 scope.go:117] "RemoveContainer" containerID="0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5" Mar 20 11:26:58 crc kubenswrapper[4846]: E0320 11:26:58.629911 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5\": container with ID starting with 0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5 not found: ID does not exist" containerID="0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.629959 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5"} err="failed to get container status \"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5\": rpc error: code = NotFound desc = could not find container \"0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5\": container with ID starting with 0c513ae1fdcf29fe46739f6a55d70cd936d1e0f9f309d997c3a81d4890c1e0a5 not found: ID does not exist" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.629993 4846 scope.go:117] "RemoveContainer" containerID="4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241" Mar 20 11:26:58 crc kubenswrapper[4846]: E0320 11:26:58.630376 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241\": container with ID starting with 4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241 not found: ID does not exist" containerID="4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.630401 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241"} err="failed to get container status \"4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241\": rpc error: code = NotFound desc = could not find container \"4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241\": container with ID starting with 4e5ae348d4bca164004dcf4f2b060ea9e0e40759d91065cb4d0c6d48618fe241 not found: ID does not exist" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.630419 4846 scope.go:117] "RemoveContainer" containerID="4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca" Mar 20 11:26:58 crc kubenswrapper[4846]: E0320 11:26:58.630762 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca\": container with ID starting with 4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca not found: ID does not exist" containerID="4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.630800 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca"} err="failed to get container status \"4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca\": rpc error: code = NotFound desc = could not find container \"4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca\": container with ID starting with 4c9a10b9afb788c9dfd2e2b12e2e2af78d5e52d2bf9b527f79da0eba1c0ad8ca not found: ID does not exist" Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.886002 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:26:58 crc kubenswrapper[4846]: I0320 11:26:58.892184 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zb6fs"] Mar 20 11:26:59 crc kubenswrapper[4846]: I0320 11:26:59.331512 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" path="/var/lib/kubelet/pods/6105d963-9e35-4d0b-897f-8b4488d105ba/volumes" Mar 20 11:27:01 crc kubenswrapper[4846]: I0320 11:27:01.322308 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:27:01 crc kubenswrapper[4846]: E0320 11:27:01.324073 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:27:12 crc kubenswrapper[4846]: I0320 11:27:12.323247 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:27:12 crc kubenswrapper[4846]: E0320 11:27:12.324564 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:27:23 crc kubenswrapper[4846]: I0320 11:27:23.322998 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:27:23 crc kubenswrapper[4846]: E0320 11:27:23.323823 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:27:37 crc kubenswrapper[4846]: I0320 11:27:37.323002 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:27:37 crc kubenswrapper[4846]: E0320 11:27:37.324035 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:27:48 crc kubenswrapper[4846]: I0320 11:27:48.322698 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:27:48 crc kubenswrapper[4846]: E0320 11:27:48.323614 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.142922 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566768-z86c4"] Mar 20 11:28:00 crc kubenswrapper[4846]: E0320 11:28:00.144165 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="extract-utilities" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.144186 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="extract-utilities" Mar 20 11:28:00 crc kubenswrapper[4846]: E0320 11:28:00.144249 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="extract-content" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.144259 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="extract-content" Mar 20 11:28:00 crc kubenswrapper[4846]: E0320 11:28:00.144271 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="registry-server" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.144279 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="registry-server" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.144457 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6105d963-9e35-4d0b-897f-8b4488d105ba" containerName="registry-server" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.145130 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.148009 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.148246 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.150647 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.151854 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566768-z86c4"] Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.280012 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhsjq\" (UniqueName: \"kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq\") pod \"auto-csr-approver-29566768-z86c4\" (UID: \"9cbd731a-e8d1-4340-93d1-8dc160c0d700\") " pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.321978 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:28:00 crc kubenswrapper[4846]: E0320 11:28:00.322198 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.381883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhsjq\" (UniqueName: \"kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq\") pod \"auto-csr-approver-29566768-z86c4\" (UID: \"9cbd731a-e8d1-4340-93d1-8dc160c0d700\") " pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.401203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhsjq\" (UniqueName: \"kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq\") pod \"auto-csr-approver-29566768-z86c4\" (UID: \"9cbd731a-e8d1-4340-93d1-8dc160c0d700\") " pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.470969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.976476 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566768-z86c4"] Mar 20 11:28:00 crc kubenswrapper[4846]: I0320 11:28:00.981965 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:28:01 crc kubenswrapper[4846]: I0320 11:28:01.055320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566768-z86c4" event={"ID":"9cbd731a-e8d1-4340-93d1-8dc160c0d700","Type":"ContainerStarted","Data":"d98b937480cb707dadcb54c345f239c77c9478a5e1ff0d416a760617b1d9ef6f"} Mar 20 11:28:04 crc kubenswrapper[4846]: I0320 11:28:04.077331 4846 generic.go:334] "Generic (PLEG): container finished" podID="9cbd731a-e8d1-4340-93d1-8dc160c0d700" containerID="abf679ca2c27ec1a0dcc4c13b3ae113116c5561e660ed246168e6bc3feb3e70d" exitCode=0 Mar 20 11:28:04 crc kubenswrapper[4846]: I0320 11:28:04.077410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566768-z86c4" event={"ID":"9cbd731a-e8d1-4340-93d1-8dc160c0d700","Type":"ContainerDied","Data":"abf679ca2c27ec1a0dcc4c13b3ae113116c5561e660ed246168e6bc3feb3e70d"} Mar 20 11:28:05 crc kubenswrapper[4846]: I0320 11:28:05.357656 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:05 crc kubenswrapper[4846]: I0320 11:28:05.482041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhsjq\" (UniqueName: \"kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq\") pod \"9cbd731a-e8d1-4340-93d1-8dc160c0d700\" (UID: \"9cbd731a-e8d1-4340-93d1-8dc160c0d700\") " Mar 20 11:28:05 crc kubenswrapper[4846]: I0320 11:28:05.489342 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq" (OuterVolumeSpecName: "kube-api-access-mhsjq") pod "9cbd731a-e8d1-4340-93d1-8dc160c0d700" (UID: "9cbd731a-e8d1-4340-93d1-8dc160c0d700"). InnerVolumeSpecName "kube-api-access-mhsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:28:05 crc kubenswrapper[4846]: I0320 11:28:05.584979 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhsjq\" (UniqueName: \"kubernetes.io/projected/9cbd731a-e8d1-4340-93d1-8dc160c0d700-kube-api-access-mhsjq\") on node \"crc\" DevicePath \"\"" Mar 20 11:28:06 crc kubenswrapper[4846]: I0320 11:28:06.100868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566768-z86c4" event={"ID":"9cbd731a-e8d1-4340-93d1-8dc160c0d700","Type":"ContainerDied","Data":"d98b937480cb707dadcb54c345f239c77c9478a5e1ff0d416a760617b1d9ef6f"} Mar 20 11:28:06 crc kubenswrapper[4846]: I0320 11:28:06.100949 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d98b937480cb707dadcb54c345f239c77c9478a5e1ff0d416a760617b1d9ef6f" Mar 20 11:28:06 crc kubenswrapper[4846]: I0320 11:28:06.100947 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566768-z86c4" Mar 20 11:28:06 crc kubenswrapper[4846]: I0320 11:28:06.437188 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566762-8l7rd"] Mar 20 11:28:06 crc kubenswrapper[4846]: I0320 11:28:06.446337 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566762-8l7rd"] Mar 20 11:28:07 crc kubenswrapper[4846]: I0320 11:28:07.332640 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f490d18-0061-48ea-aa5c-dc0e5118e028" path="/var/lib/kubelet/pods/8f490d18-0061-48ea-aa5c-dc0e5118e028/volumes" Mar 20 11:28:11 crc kubenswrapper[4846]: I0320 11:28:11.322996 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:28:11 crc kubenswrapper[4846]: E0320 11:28:11.324090 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:28:14 crc kubenswrapper[4846]: I0320 11:28:14.695561 4846 scope.go:117] "RemoveContainer" containerID="a3be040ae25d0267a70d7ea393eb534cd49b73733c6b684ce5f3de4e8f674880" Mar 20 11:28:24 crc kubenswrapper[4846]: I0320 11:28:24.323873 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:28:24 crc kubenswrapper[4846]: E0320 11:28:24.325120 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:28:38 crc kubenswrapper[4846]: I0320 11:28:38.323368 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:28:38 crc kubenswrapper[4846]: E0320 11:28:38.324515 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:28:50 crc kubenswrapper[4846]: I0320 11:28:50.323231 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:28:50 crc kubenswrapper[4846]: E0320 11:28:50.324159 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:29:04 crc kubenswrapper[4846]: I0320 11:29:04.322604 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:29:04 crc kubenswrapper[4846]: E0320 11:29:04.323722 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:29:15 crc kubenswrapper[4846]: I0320 11:29:15.323468 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:29:15 crc kubenswrapper[4846]: I0320 11:29:15.739653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e"} Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.156449 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566770-hq99r"] Mar 20 11:30:00 crc kubenswrapper[4846]: E0320 11:30:00.157817 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbd731a-e8d1-4340-93d1-8dc160c0d700" containerName="oc" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.157836 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbd731a-e8d1-4340-93d1-8dc160c0d700" containerName="oc" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.158067 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbd731a-e8d1-4340-93d1-8dc160c0d700" containerName="oc" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.158697 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.170133 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.171197 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.172087 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.193354 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r"] Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.196234 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.201587 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.201943 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.203717 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566770-hq99r"] Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.218634 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r"] Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.306598 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.306824 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7r89\" (UniqueName: \"kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.306914 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqcm4\" (UniqueName: \"kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4\") pod \"auto-csr-approver-29566770-hq99r\" (UID: \"76bc9120-ef21-47ca-ab65-8855bea9dffc\") " pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.306994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.409249 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.409327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7r89\" (UniqueName: \"kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.409358 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqcm4\" (UniqueName: \"kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4\") pod \"auto-csr-approver-29566770-hq99r\" (UID: \"76bc9120-ef21-47ca-ab65-8855bea9dffc\") " pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.409387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.410592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.416990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.429104 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqcm4\" (UniqueName: \"kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4\") pod \"auto-csr-approver-29566770-hq99r\" (UID: \"76bc9120-ef21-47ca-ab65-8855bea9dffc\") " pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.434624 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7r89\" (UniqueName: \"kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89\") pod \"collect-profiles-29566770-qzl5r\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.492041 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.525179 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:00 crc kubenswrapper[4846]: I0320 11:30:00.964117 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566770-hq99r"] Mar 20 11:30:01 crc kubenswrapper[4846]: I0320 11:30:01.038932 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r"] Mar 20 11:30:01 crc kubenswrapper[4846]: W0320 11:30:01.041125 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fbb0dd_0dcf_4d65_830a_7a8fc841303f.slice/crio-137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e WatchSource:0}: Error finding container 137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e: Status 404 returned error can't find the container with id 137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e Mar 20 11:30:01 crc kubenswrapper[4846]: I0320 11:30:01.123054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" event={"ID":"63fbb0dd-0dcf-4d65-830a-7a8fc841303f","Type":"ContainerStarted","Data":"137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e"} Mar 20 11:30:01 crc kubenswrapper[4846]: I0320 11:30:01.125970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566770-hq99r" event={"ID":"76bc9120-ef21-47ca-ab65-8855bea9dffc","Type":"ContainerStarted","Data":"aea22f2150cd222578de5f1edce4eef750b3b1d0747b951bf57017e4eb0612db"} Mar 20 11:30:02 crc kubenswrapper[4846]: I0320 11:30:02.134850 4846 generic.go:334] "Generic (PLEG): container finished" podID="63fbb0dd-0dcf-4d65-830a-7a8fc841303f" containerID="e83fd9ff9ae6abb1deeae7a08de64d0d35d59655fe863c73d12959a3e932207f" exitCode=0 Mar 20 11:30:02 crc kubenswrapper[4846]: I0320 11:30:02.134919 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" event={"ID":"63fbb0dd-0dcf-4d65-830a-7a8fc841303f","Type":"ContainerDied","Data":"e83fd9ff9ae6abb1deeae7a08de64d0d35d59655fe863c73d12959a3e932207f"} Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.483543 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.599716 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume\") pod \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.599890 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume\") pod \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.599956 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7r89\" (UniqueName: \"kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89\") pod \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\" (UID: \"63fbb0dd-0dcf-4d65-830a-7a8fc841303f\") " Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.601300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume" (OuterVolumeSpecName: "config-volume") pod "63fbb0dd-0dcf-4d65-830a-7a8fc841303f" (UID: "63fbb0dd-0dcf-4d65-830a-7a8fc841303f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.610000 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89" (OuterVolumeSpecName: "kube-api-access-f7r89") pod "63fbb0dd-0dcf-4d65-830a-7a8fc841303f" (UID: "63fbb0dd-0dcf-4d65-830a-7a8fc841303f"). InnerVolumeSpecName "kube-api-access-f7r89". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.615031 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "63fbb0dd-0dcf-4d65-830a-7a8fc841303f" (UID: "63fbb0dd-0dcf-4d65-830a-7a8fc841303f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.701952 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.703196 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:30:03 crc kubenswrapper[4846]: I0320 11:30:03.703265 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7r89\" (UniqueName: \"kubernetes.io/projected/63fbb0dd-0dcf-4d65-830a-7a8fc841303f-kube-api-access-f7r89\") on node \"crc\" DevicePath \"\"" Mar 20 11:30:04 crc kubenswrapper[4846]: I0320 11:30:04.151843 4846 generic.go:334] "Generic (PLEG): container finished" podID="76bc9120-ef21-47ca-ab65-8855bea9dffc" containerID="1ecb51733fa176a95c9983212d9cac23d49ce05d33b89368fdcc0f9bc7283931" exitCode=0 Mar 20 11:30:04 crc kubenswrapper[4846]: I0320 11:30:04.152175 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566770-hq99r" event={"ID":"76bc9120-ef21-47ca-ab65-8855bea9dffc","Type":"ContainerDied","Data":"1ecb51733fa176a95c9983212d9cac23d49ce05d33b89368fdcc0f9bc7283931"} Mar 20 11:30:04 crc kubenswrapper[4846]: I0320 11:30:04.154268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" event={"ID":"63fbb0dd-0dcf-4d65-830a-7a8fc841303f","Type":"ContainerDied","Data":"137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e"} Mar 20 11:30:04 crc kubenswrapper[4846]: I0320 11:30:04.154302 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="137f788ea6d35e0ef9eac768557cf8bfacbaf9e41f9d1a767bbb69177570da0e" Mar 20 11:30:04 crc kubenswrapper[4846]: I0320 11:30:04.154459 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566770-qzl5r" Mar 20 11:30:05 crc kubenswrapper[4846]: I0320 11:30:05.493759 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:05 crc kubenswrapper[4846]: I0320 11:30:05.638887 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqcm4\" (UniqueName: \"kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4\") pod \"76bc9120-ef21-47ca-ab65-8855bea9dffc\" (UID: \"76bc9120-ef21-47ca-ab65-8855bea9dffc\") " Mar 20 11:30:05 crc kubenswrapper[4846]: I0320 11:30:05.646705 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4" (OuterVolumeSpecName: "kube-api-access-tqcm4") pod "76bc9120-ef21-47ca-ab65-8855bea9dffc" (UID: "76bc9120-ef21-47ca-ab65-8855bea9dffc"). InnerVolumeSpecName "kube-api-access-tqcm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:30:05 crc kubenswrapper[4846]: I0320 11:30:05.740382 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqcm4\" (UniqueName: \"kubernetes.io/projected/76bc9120-ef21-47ca-ab65-8855bea9dffc-kube-api-access-tqcm4\") on node \"crc\" DevicePath \"\"" Mar 20 11:30:06 crc kubenswrapper[4846]: I0320 11:30:06.171232 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566770-hq99r" event={"ID":"76bc9120-ef21-47ca-ab65-8855bea9dffc","Type":"ContainerDied","Data":"aea22f2150cd222578de5f1edce4eef750b3b1d0747b951bf57017e4eb0612db"} Mar 20 11:30:06 crc kubenswrapper[4846]: I0320 11:30:06.171278 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aea22f2150cd222578de5f1edce4eef750b3b1d0747b951bf57017e4eb0612db" Mar 20 11:30:06 crc kubenswrapper[4846]: I0320 11:30:06.171371 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566770-hq99r" Mar 20 11:30:06 crc kubenswrapper[4846]: I0320 11:30:06.577300 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566764-5ngks"] Mar 20 11:30:06 crc kubenswrapper[4846]: I0320 11:30:06.588698 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566764-5ngks"] Mar 20 11:30:07 crc kubenswrapper[4846]: I0320 11:30:07.331592 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b73c17a-882b-4565-bc01-ddc1f6576c0e" path="/var/lib/kubelet/pods/0b73c17a-882b-4565-bc01-ddc1f6576c0e/volumes" Mar 20 11:30:14 crc kubenswrapper[4846]: I0320 11:30:14.785353 4846 scope.go:117] "RemoveContainer" containerID="cbeec4103f0bc3a35efe6d52b80ba24cc8271da7cf1e9d5f7c630e26e97d6abc" Mar 20 11:31:39 crc kubenswrapper[4846]: I0320 11:31:39.677922 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:31:39 crc kubenswrapper[4846]: I0320 11:31:39.678711 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.163274 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:31:43 crc kubenswrapper[4846]: E0320 11:31:43.164175 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fbb0dd-0dcf-4d65-830a-7a8fc841303f" containerName="collect-profiles" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.164193 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fbb0dd-0dcf-4d65-830a-7a8fc841303f" containerName="collect-profiles" Mar 20 11:31:43 crc kubenswrapper[4846]: E0320 11:31:43.164209 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76bc9120-ef21-47ca-ab65-8855bea9dffc" containerName="oc" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.164216 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76bc9120-ef21-47ca-ab65-8855bea9dffc" containerName="oc" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.164376 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="76bc9120-ef21-47ca-ab65-8855bea9dffc" containerName="oc" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.164392 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fbb0dd-0dcf-4d65-830a-7a8fc841303f" containerName="collect-profiles" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.166061 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.187564 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.221429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.221682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.221818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwfk5\" (UniqueName: \"kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.323042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.323118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.323179 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwfk5\" (UniqueName: \"kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.323724 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.323809 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.351977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwfk5\" (UniqueName: \"kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5\") pod \"redhat-operators-6z9xw\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.489377 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:43 crc kubenswrapper[4846]: I0320 11:31:43.995882 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:31:44 crc kubenswrapper[4846]: I0320 11:31:44.982581 4846 generic.go:334] "Generic (PLEG): container finished" podID="9263e866-bf8c-40e1-bdff-804c53158431" containerID="afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5" exitCode=0 Mar 20 11:31:44 crc kubenswrapper[4846]: I0320 11:31:44.982650 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerDied","Data":"afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5"} Mar 20 11:31:44 crc kubenswrapper[4846]: I0320 11:31:44.983016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerStarted","Data":"04f4186409b909b777038b9c516f369efa8d8dbab367ac3c1156b5523f68bebf"} Mar 20 11:31:47 crc kubenswrapper[4846]: I0320 11:31:47.004255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerStarted","Data":"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f"} Mar 20 11:31:48 crc kubenswrapper[4846]: I0320 11:31:48.019130 4846 generic.go:334] "Generic (PLEG): container finished" podID="9263e866-bf8c-40e1-bdff-804c53158431" containerID="d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f" exitCode=0 Mar 20 11:31:48 crc kubenswrapper[4846]: I0320 11:31:48.019218 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerDied","Data":"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f"} Mar 20 11:31:49 crc kubenswrapper[4846]: I0320 11:31:49.029245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerStarted","Data":"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e"} Mar 20 11:31:49 crc kubenswrapper[4846]: I0320 11:31:49.052927 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6z9xw" podStartSLOduration=2.587018585 podStartE2EDuration="6.052876777s" podCreationTimestamp="2026-03-20 11:31:43 +0000 UTC" firstStartedPulling="2026-03-20 11:31:44.985317166 +0000 UTC m=+2076.351725402" lastFinishedPulling="2026-03-20 11:31:48.451175338 +0000 UTC m=+2079.817583594" observedRunningTime="2026-03-20 11:31:49.050745925 +0000 UTC m=+2080.417154161" watchObservedRunningTime="2026-03-20 11:31:49.052876777 +0000 UTC m=+2080.419285033" Mar 20 11:31:53 crc kubenswrapper[4846]: I0320 11:31:53.490054 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:53 crc kubenswrapper[4846]: I0320 11:31:53.491027 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:31:54 crc kubenswrapper[4846]: I0320 11:31:54.532858 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6z9xw" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="registry-server" probeResult="failure" output=< Mar 20 11:31:54 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Mar 20 11:31:54 crc kubenswrapper[4846]: > Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.143145 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566772-2zq6x"] Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.144869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.150278 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.150957 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.151135 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.160299 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566772-2zq6x"] Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.307481 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp59v\" (UniqueName: \"kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v\") pod \"auto-csr-approver-29566772-2zq6x\" (UID: \"1de333c6-d33e-41e2-bf7e-2c9a137e37c3\") " pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.409807 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp59v\" (UniqueName: \"kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v\") pod \"auto-csr-approver-29566772-2zq6x\" (UID: \"1de333c6-d33e-41e2-bf7e-2c9a137e37c3\") " pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.435595 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp59v\" (UniqueName: \"kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v\") pod \"auto-csr-approver-29566772-2zq6x\" (UID: \"1de333c6-d33e-41e2-bf7e-2c9a137e37c3\") " pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.472582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:00 crc kubenswrapper[4846]: I0320 11:32:00.930332 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566772-2zq6x"] Mar 20 11:32:00 crc kubenswrapper[4846]: W0320 11:32:00.936080 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1de333c6_d33e_41e2_bf7e_2c9a137e37c3.slice/crio-cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed WatchSource:0}: Error finding container cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed: Status 404 returned error can't find the container with id cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed Mar 20 11:32:01 crc kubenswrapper[4846]: I0320 11:32:01.127524 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" event={"ID":"1de333c6-d33e-41e2-bf7e-2c9a137e37c3","Type":"ContainerStarted","Data":"cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed"} Mar 20 11:32:03 crc kubenswrapper[4846]: I0320 11:32:03.151939 4846 generic.go:334] "Generic (PLEG): container finished" podID="1de333c6-d33e-41e2-bf7e-2c9a137e37c3" containerID="7ee2fc2b5ed323fcb85163f190103adc7135fbcf245bb405a07efa7bf509335d" exitCode=0 Mar 20 11:32:03 crc kubenswrapper[4846]: I0320 11:32:03.152566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" event={"ID":"1de333c6-d33e-41e2-bf7e-2c9a137e37c3","Type":"ContainerDied","Data":"7ee2fc2b5ed323fcb85163f190103adc7135fbcf245bb405a07efa7bf509335d"} Mar 20 11:32:03 crc kubenswrapper[4846]: I0320 11:32:03.534652 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:32:03 crc kubenswrapper[4846]: I0320 11:32:03.616343 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:32:03 crc kubenswrapper[4846]: I0320 11:32:03.774945 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:32:04 crc kubenswrapper[4846]: I0320 11:32:04.470703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:04 crc kubenswrapper[4846]: I0320 11:32:04.581404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp59v\" (UniqueName: \"kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v\") pod \"1de333c6-d33e-41e2-bf7e-2c9a137e37c3\" (UID: \"1de333c6-d33e-41e2-bf7e-2c9a137e37c3\") " Mar 20 11:32:04 crc kubenswrapper[4846]: I0320 11:32:04.590395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v" (OuterVolumeSpecName: "kube-api-access-pp59v") pod "1de333c6-d33e-41e2-bf7e-2c9a137e37c3" (UID: "1de333c6-d33e-41e2-bf7e-2c9a137e37c3"). InnerVolumeSpecName "kube-api-access-pp59v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:32:04 crc kubenswrapper[4846]: I0320 11:32:04.683887 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp59v\" (UniqueName: \"kubernetes.io/projected/1de333c6-d33e-41e2-bf7e-2c9a137e37c3-kube-api-access-pp59v\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.175991 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6z9xw" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="registry-server" containerID="cri-o://e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e" gracePeriod=2 Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.176138 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.177377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566772-2zq6x" event={"ID":"1de333c6-d33e-41e2-bf7e-2c9a137e37c3","Type":"ContainerDied","Data":"cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed"} Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.177460 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb2594f98f03275cfc213df5c52458bf1d3443a7dd9125e4c1feeb2f04bcd4ed" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.567368 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566766-jsqld"] Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.577548 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566766-jsqld"] Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.599820 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.704536 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content\") pod \"9263e866-bf8c-40e1-bdff-804c53158431\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.704703 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities\") pod \"9263e866-bf8c-40e1-bdff-804c53158431\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.704846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwfk5\" (UniqueName: \"kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5\") pod \"9263e866-bf8c-40e1-bdff-804c53158431\" (UID: \"9263e866-bf8c-40e1-bdff-804c53158431\") " Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.706368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities" (OuterVolumeSpecName: "utilities") pod "9263e866-bf8c-40e1-bdff-804c53158431" (UID: "9263e866-bf8c-40e1-bdff-804c53158431"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.728491 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5" (OuterVolumeSpecName: "kube-api-access-mwfk5") pod "9263e866-bf8c-40e1-bdff-804c53158431" (UID: "9263e866-bf8c-40e1-bdff-804c53158431"). InnerVolumeSpecName "kube-api-access-mwfk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.806679 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwfk5\" (UniqueName: \"kubernetes.io/projected/9263e866-bf8c-40e1-bdff-804c53158431-kube-api-access-mwfk5\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.806721 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.837971 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9263e866-bf8c-40e1-bdff-804c53158431" (UID: "9263e866-bf8c-40e1-bdff-804c53158431"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:32:05 crc kubenswrapper[4846]: I0320 11:32:05.912558 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9263e866-bf8c-40e1-bdff-804c53158431-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.187268 4846 generic.go:334] "Generic (PLEG): container finished" podID="9263e866-bf8c-40e1-bdff-804c53158431" containerID="e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e" exitCode=0 Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.187328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerDied","Data":"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e"} Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.187343 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6z9xw" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.187379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6z9xw" event={"ID":"9263e866-bf8c-40e1-bdff-804c53158431","Type":"ContainerDied","Data":"04f4186409b909b777038b9c516f369efa8d8dbab367ac3c1156b5523f68bebf"} Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.187413 4846 scope.go:117] "RemoveContainer" containerID="e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.214505 4846 scope.go:117] "RemoveContainer" containerID="d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.227357 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.239383 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6z9xw"] Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.248410 4846 scope.go:117] "RemoveContainer" containerID="afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.266087 4846 scope.go:117] "RemoveContainer" containerID="e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e" Mar 20 11:32:06 crc kubenswrapper[4846]: E0320 11:32:06.266853 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e\": container with ID starting with e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e not found: ID does not exist" containerID="e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.266955 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e"} err="failed to get container status \"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e\": rpc error: code = NotFound desc = could not find container \"e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e\": container with ID starting with e2a00854ede593e2f9eb28d660cc5416de5defc2c46922ee72a6041b5572f66e not found: ID does not exist" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.267013 4846 scope.go:117] "RemoveContainer" containerID="d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f" Mar 20 11:32:06 crc kubenswrapper[4846]: E0320 11:32:06.267666 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f\": container with ID starting with d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f not found: ID does not exist" containerID="d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.267746 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f"} err="failed to get container status \"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f\": rpc error: code = NotFound desc = could not find container \"d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f\": container with ID starting with d8d7470efd42473f3c4d7596f555d0003d71da47a03dbad3efe725a69a8cbb5f not found: ID does not exist" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.267795 4846 scope.go:117] "RemoveContainer" containerID="afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5" Mar 20 11:32:06 crc kubenswrapper[4846]: E0320 11:32:06.268299 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5\": container with ID starting with afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5 not found: ID does not exist" containerID="afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5" Mar 20 11:32:06 crc kubenswrapper[4846]: I0320 11:32:06.268340 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5"} err="failed to get container status \"afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5\": rpc error: code = NotFound desc = could not find container \"afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5\": container with ID starting with afd6e6b92813f076879f3d85d7487cfe8de70c99fc589d1d9ef75a323f9a07f5 not found: ID does not exist" Mar 20 11:32:07 crc kubenswrapper[4846]: I0320 11:32:07.335438 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9263e866-bf8c-40e1-bdff-804c53158431" path="/var/lib/kubelet/pods/9263e866-bf8c-40e1-bdff-804c53158431/volumes" Mar 20 11:32:07 crc kubenswrapper[4846]: I0320 11:32:07.336692 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd79e98-23c2-451f-91a6-8bc35bd6380f" path="/var/lib/kubelet/pods/9dd79e98-23c2-451f-91a6-8bc35bd6380f/volumes" Mar 20 11:32:09 crc kubenswrapper[4846]: I0320 11:32:09.677945 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:32:09 crc kubenswrapper[4846]: I0320 11:32:09.678050 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:32:14 crc kubenswrapper[4846]: I0320 11:32:14.895375 4846 scope.go:117] "RemoveContainer" containerID="cb5e084d7f0ee267eb11ba352ed35f56d4ba6ff7c7dfb1951d987e2d534a02aa" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.884297 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:34 crc kubenswrapper[4846]: E0320 11:32:34.885391 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="extract-content" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885412 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="extract-content" Mar 20 11:32:34 crc kubenswrapper[4846]: E0320 11:32:34.885431 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de333c6-d33e-41e2-bf7e-2c9a137e37c3" containerName="oc" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885440 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de333c6-d33e-41e2-bf7e-2c9a137e37c3" containerName="oc" Mar 20 11:32:34 crc kubenswrapper[4846]: E0320 11:32:34.885454 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="registry-server" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885462 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="registry-server" Mar 20 11:32:34 crc kubenswrapper[4846]: E0320 11:32:34.885475 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="extract-utilities" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885482 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="extract-utilities" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885649 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de333c6-d33e-41e2-bf7e-2c9a137e37c3" containerName="oc" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.885676 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9263e866-bf8c-40e1-bdff-804c53158431" containerName="registry-server" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.886978 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:34 crc kubenswrapper[4846]: I0320 11:32:34.907632 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.042077 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.042564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52mch\" (UniqueName: \"kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.042593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.143796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52mch\" (UniqueName: \"kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.143857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.143937 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.144447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.144633 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.167043 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52mch\" (UniqueName: \"kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch\") pod \"redhat-marketplace-glhcg\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.219047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:35 crc kubenswrapper[4846]: I0320 11:32:35.685786 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:36 crc kubenswrapper[4846]: I0320 11:32:36.478797 4846 generic.go:334] "Generic (PLEG): container finished" podID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerID="f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da" exitCode=0 Mar 20 11:32:36 crc kubenswrapper[4846]: I0320 11:32:36.479318 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerDied","Data":"f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da"} Mar 20 11:32:36 crc kubenswrapper[4846]: I0320 11:32:36.479374 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerStarted","Data":"0c76d0e77028d7248a3db5e71ac186ccc977991182d352265ee7aea5a3fcf79a"} Mar 20 11:32:37 crc kubenswrapper[4846]: I0320 11:32:37.491410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerStarted","Data":"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c"} Mar 20 11:32:38 crc kubenswrapper[4846]: I0320 11:32:38.502368 4846 generic.go:334] "Generic (PLEG): container finished" podID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerID="1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c" exitCode=0 Mar 20 11:32:38 crc kubenswrapper[4846]: I0320 11:32:38.502482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerDied","Data":"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c"} Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.512214 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerStarted","Data":"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7"} Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.537851 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-glhcg" podStartSLOduration=3.057616592 podStartE2EDuration="5.537825475s" podCreationTimestamp="2026-03-20 11:32:34 +0000 UTC" firstStartedPulling="2026-03-20 11:32:36.484520598 +0000 UTC m=+2127.850928844" lastFinishedPulling="2026-03-20 11:32:38.964729491 +0000 UTC m=+2130.331137727" observedRunningTime="2026-03-20 11:32:39.533042488 +0000 UTC m=+2130.899450744" watchObservedRunningTime="2026-03-20 11:32:39.537825475 +0000 UTC m=+2130.904233731" Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.678422 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.678510 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.678575 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.679424 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:32:39 crc kubenswrapper[4846]: I0320 11:32:39.679503 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e" gracePeriod=600 Mar 20 11:32:40 crc kubenswrapper[4846]: I0320 11:32:40.522061 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e" exitCode=0 Mar 20 11:32:40 crc kubenswrapper[4846]: I0320 11:32:40.524023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e"} Mar 20 11:32:40 crc kubenswrapper[4846]: I0320 11:32:40.524091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72"} Mar 20 11:32:40 crc kubenswrapper[4846]: I0320 11:32:40.524119 4846 scope.go:117] "RemoveContainer" containerID="1de3531c2e90edf29cb13460136cb81e1221665663c4821cff6ca10d56cb10b2" Mar 20 11:32:45 crc kubenswrapper[4846]: I0320 11:32:45.220040 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:45 crc kubenswrapper[4846]: I0320 11:32:45.220867 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:45 crc kubenswrapper[4846]: I0320 11:32:45.280372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:45 crc kubenswrapper[4846]: I0320 11:32:45.621879 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:45 crc kubenswrapper[4846]: I0320 11:32:45.680344 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:47 crc kubenswrapper[4846]: I0320 11:32:47.587773 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-glhcg" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="registry-server" containerID="cri-o://c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7" gracePeriod=2 Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.054022 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.165615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities\") pod \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.165837 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content\") pod \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.165963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52mch\" (UniqueName: \"kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch\") pod \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\" (UID: \"437ca1f7-46f6-46bb-b1e2-44f5898cb194\") " Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.167087 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities" (OuterVolumeSpecName: "utilities") pod "437ca1f7-46f6-46bb-b1e2-44f5898cb194" (UID: "437ca1f7-46f6-46bb-b1e2-44f5898cb194"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.174159 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch" (OuterVolumeSpecName: "kube-api-access-52mch") pod "437ca1f7-46f6-46bb-b1e2-44f5898cb194" (UID: "437ca1f7-46f6-46bb-b1e2-44f5898cb194"). InnerVolumeSpecName "kube-api-access-52mch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.268247 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52mch\" (UniqueName: \"kubernetes.io/projected/437ca1f7-46f6-46bb-b1e2-44f5898cb194-kube-api-access-52mch\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.268370 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.313964 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "437ca1f7-46f6-46bb-b1e2-44f5898cb194" (UID: "437ca1f7-46f6-46bb-b1e2-44f5898cb194"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.369660 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/437ca1f7-46f6-46bb-b1e2-44f5898cb194-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.603161 4846 generic.go:334] "Generic (PLEG): container finished" podID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerID="c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7" exitCode=0 Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.603253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerDied","Data":"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7"} Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.603303 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glhcg" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.603321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glhcg" event={"ID":"437ca1f7-46f6-46bb-b1e2-44f5898cb194","Type":"ContainerDied","Data":"0c76d0e77028d7248a3db5e71ac186ccc977991182d352265ee7aea5a3fcf79a"} Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.603338 4846 scope.go:117] "RemoveContainer" containerID="c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.633181 4846 scope.go:117] "RemoveContainer" containerID="1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.650477 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.658338 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-glhcg"] Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.663635 4846 scope.go:117] "RemoveContainer" containerID="f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.718194 4846 scope.go:117] "RemoveContainer" containerID="c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7" Mar 20 11:32:48 crc kubenswrapper[4846]: E0320 11:32:48.719272 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7\": container with ID starting with c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7 not found: ID does not exist" containerID="c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.719334 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7"} err="failed to get container status \"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7\": rpc error: code = NotFound desc = could not find container \"c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7\": container with ID starting with c31f7317501fc4effc092bbd0c876f0bcab51bb33437a5085e3e26a68b09beb7 not found: ID does not exist" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.719367 4846 scope.go:117] "RemoveContainer" containerID="1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c" Mar 20 11:32:48 crc kubenswrapper[4846]: E0320 11:32:48.720076 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c\": container with ID starting with 1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c not found: ID does not exist" containerID="1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.720123 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c"} err="failed to get container status \"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c\": rpc error: code = NotFound desc = could not find container \"1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c\": container with ID starting with 1216e763fe82378b6b6368e8315675a26b2a942e7e02e5312bd97b2bc4d9009c not found: ID does not exist" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.720156 4846 scope.go:117] "RemoveContainer" containerID="f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da" Mar 20 11:32:48 crc kubenswrapper[4846]: E0320 11:32:48.720720 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da\": container with ID starting with f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da not found: ID does not exist" containerID="f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da" Mar 20 11:32:48 crc kubenswrapper[4846]: I0320 11:32:48.720747 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da"} err="failed to get container status \"f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da\": rpc error: code = NotFound desc = could not find container \"f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da\": container with ID starting with f2de3f063dd4248ee1f5f45e49107bb60badbe00ef207cf72e4243eea5e655da not found: ID does not exist" Mar 20 11:32:49 crc kubenswrapper[4846]: I0320 11:32:49.336885 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" path="/var/lib/kubelet/pods/437ca1f7-46f6-46bb-b1e2-44f5898cb194/volumes" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.146611 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566774-thkdc"] Mar 20 11:34:00 crc kubenswrapper[4846]: E0320 11:34:00.147874 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="extract-utilities" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.147901 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="extract-utilities" Mar 20 11:34:00 crc kubenswrapper[4846]: E0320 11:34:00.147936 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="extract-content" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.147944 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="extract-content" Mar 20 11:34:00 crc kubenswrapper[4846]: E0320 11:34:00.147967 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="registry-server" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.147979 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="registry-server" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.148173 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="437ca1f7-46f6-46bb-b1e2-44f5898cb194" containerName="registry-server" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.148857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.161377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566774-thkdc"] Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.171179 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.175191 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.175966 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.237435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ns75\" (UniqueName: \"kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75\") pod \"auto-csr-approver-29566774-thkdc\" (UID: \"d147a127-88f4-42bb-9879-5f2ff1ace4eb\") " pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.338862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ns75\" (UniqueName: \"kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75\") pod \"auto-csr-approver-29566774-thkdc\" (UID: \"d147a127-88f4-42bb-9879-5f2ff1ace4eb\") " pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.363177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ns75\" (UniqueName: \"kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75\") pod \"auto-csr-approver-29566774-thkdc\" (UID: \"d147a127-88f4-42bb-9879-5f2ff1ace4eb\") " pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.482510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.920055 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566774-thkdc"] Mar 20 11:34:00 crc kubenswrapper[4846]: I0320 11:34:00.932306 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:34:01 crc kubenswrapper[4846]: I0320 11:34:01.220297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566774-thkdc" event={"ID":"d147a127-88f4-42bb-9879-5f2ff1ace4eb","Type":"ContainerStarted","Data":"0c8caaaafbab5e4ea48930899079fbe7a15adca69526bf90ae0ce96a4d84d5cf"} Mar 20 11:34:02 crc kubenswrapper[4846]: I0320 11:34:02.230319 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566774-thkdc" event={"ID":"d147a127-88f4-42bb-9879-5f2ff1ace4eb","Type":"ContainerStarted","Data":"f40820e0af62d8586713fc6eb7bff38f6a48bec8a116c797b2099dfb160a3ab5"} Mar 20 11:34:02 crc kubenswrapper[4846]: I0320 11:34:02.250130 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566774-thkdc" podStartSLOduration=1.213327066 podStartE2EDuration="2.250109063s" podCreationTimestamp="2026-03-20 11:34:00 +0000 UTC" firstStartedPulling="2026-03-20 11:34:00.932018805 +0000 UTC m=+2212.298427041" lastFinishedPulling="2026-03-20 11:34:01.968800802 +0000 UTC m=+2213.335209038" observedRunningTime="2026-03-20 11:34:02.246942914 +0000 UTC m=+2213.613351170" watchObservedRunningTime="2026-03-20 11:34:02.250109063 +0000 UTC m=+2213.616517309" Mar 20 11:34:03 crc kubenswrapper[4846]: I0320 11:34:03.240848 4846 generic.go:334] "Generic (PLEG): container finished" podID="d147a127-88f4-42bb-9879-5f2ff1ace4eb" containerID="f40820e0af62d8586713fc6eb7bff38f6a48bec8a116c797b2099dfb160a3ab5" exitCode=0 Mar 20 11:34:03 crc kubenswrapper[4846]: I0320 11:34:03.240971 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566774-thkdc" event={"ID":"d147a127-88f4-42bb-9879-5f2ff1ace4eb","Type":"ContainerDied","Data":"f40820e0af62d8586713fc6eb7bff38f6a48bec8a116c797b2099dfb160a3ab5"} Mar 20 11:34:04 crc kubenswrapper[4846]: I0320 11:34:04.605929 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:04 crc kubenswrapper[4846]: I0320 11:34:04.710963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ns75\" (UniqueName: \"kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75\") pod \"d147a127-88f4-42bb-9879-5f2ff1ace4eb\" (UID: \"d147a127-88f4-42bb-9879-5f2ff1ace4eb\") " Mar 20 11:34:04 crc kubenswrapper[4846]: I0320 11:34:04.719574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75" (OuterVolumeSpecName: "kube-api-access-7ns75") pod "d147a127-88f4-42bb-9879-5f2ff1ace4eb" (UID: "d147a127-88f4-42bb-9879-5f2ff1ace4eb"). InnerVolumeSpecName "kube-api-access-7ns75". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:34:04 crc kubenswrapper[4846]: I0320 11:34:04.812509 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ns75\" (UniqueName: \"kubernetes.io/projected/d147a127-88f4-42bb-9879-5f2ff1ace4eb-kube-api-access-7ns75\") on node \"crc\" DevicePath \"\"" Mar 20 11:34:05 crc kubenswrapper[4846]: I0320 11:34:05.263729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566774-thkdc" event={"ID":"d147a127-88f4-42bb-9879-5f2ff1ace4eb","Type":"ContainerDied","Data":"0c8caaaafbab5e4ea48930899079fbe7a15adca69526bf90ae0ce96a4d84d5cf"} Mar 20 11:34:05 crc kubenswrapper[4846]: I0320 11:34:05.263796 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c8caaaafbab5e4ea48930899079fbe7a15adca69526bf90ae0ce96a4d84d5cf" Mar 20 11:34:05 crc kubenswrapper[4846]: I0320 11:34:05.264176 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566774-thkdc" Mar 20 11:34:05 crc kubenswrapper[4846]: I0320 11:34:05.341344 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566768-z86c4"] Mar 20 11:34:05 crc kubenswrapper[4846]: I0320 11:34:05.341406 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566768-z86c4"] Mar 20 11:34:07 crc kubenswrapper[4846]: I0320 11:34:07.333416 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbd731a-e8d1-4340-93d1-8dc160c0d700" path="/var/lib/kubelet/pods/9cbd731a-e8d1-4340-93d1-8dc160c0d700/volumes" Mar 20 11:34:15 crc kubenswrapper[4846]: I0320 11:34:15.010722 4846 scope.go:117] "RemoveContainer" containerID="abf679ca2c27ec1a0dcc4c13b3ae113116c5561e660ed246168e6bc3feb3e70d" Mar 20 11:35:09 crc kubenswrapper[4846]: I0320 11:35:09.677838 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:35:09 crc kubenswrapper[4846]: I0320 11:35:09.678769 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:35:39 crc kubenswrapper[4846]: I0320 11:35:39.677803 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:35:39 crc kubenswrapper[4846]: I0320 11:35:39.678723 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.152090 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566776-jc9wx"] Mar 20 11:36:00 crc kubenswrapper[4846]: E0320 11:36:00.154585 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d147a127-88f4-42bb-9879-5f2ff1ace4eb" containerName="oc" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.154678 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d147a127-88f4-42bb-9879-5f2ff1ace4eb" containerName="oc" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.154964 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d147a127-88f4-42bb-9879-5f2ff1ace4eb" containerName="oc" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.155671 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.158861 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.159184 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.159295 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.173968 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566776-jc9wx"] Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.234720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86xwf\" (UniqueName: \"kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf\") pod \"auto-csr-approver-29566776-jc9wx\" (UID: \"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6\") " pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.337945 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86xwf\" (UniqueName: \"kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf\") pod \"auto-csr-approver-29566776-jc9wx\" (UID: \"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6\") " pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.359133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86xwf\" (UniqueName: \"kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf\") pod \"auto-csr-approver-29566776-jc9wx\" (UID: \"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6\") " pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.507661 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:00 crc kubenswrapper[4846]: I0320 11:36:00.960662 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566776-jc9wx"] Mar 20 11:36:01 crc kubenswrapper[4846]: I0320 11:36:01.152260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" event={"ID":"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6","Type":"ContainerStarted","Data":"42cc748eef09e4d65c9df2e952d44f17290e20653f67c20c594f3d33bb5422f7"} Mar 20 11:36:03 crc kubenswrapper[4846]: I0320 11:36:03.170261 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" containerID="8b886ad83c6a569bc39cb8dd3efb39bdac5c959c39b0d126de1699cf04253bcb" exitCode=0 Mar 20 11:36:03 crc kubenswrapper[4846]: I0320 11:36:03.170428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" event={"ID":"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6","Type":"ContainerDied","Data":"8b886ad83c6a569bc39cb8dd3efb39bdac5c959c39b0d126de1699cf04253bcb"} Mar 20 11:36:04 crc kubenswrapper[4846]: I0320 11:36:04.505571 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:04 crc kubenswrapper[4846]: I0320 11:36:04.604204 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86xwf\" (UniqueName: \"kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf\") pod \"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6\" (UID: \"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6\") " Mar 20 11:36:04 crc kubenswrapper[4846]: I0320 11:36:04.612093 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf" (OuterVolumeSpecName: "kube-api-access-86xwf") pod "8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" (UID: "8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6"). InnerVolumeSpecName "kube-api-access-86xwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:36:04 crc kubenswrapper[4846]: I0320 11:36:04.706767 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86xwf\" (UniqueName: \"kubernetes.io/projected/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6-kube-api-access-86xwf\") on node \"crc\" DevicePath \"\"" Mar 20 11:36:05 crc kubenswrapper[4846]: I0320 11:36:05.197244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" event={"ID":"8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6","Type":"ContainerDied","Data":"42cc748eef09e4d65c9df2e952d44f17290e20653f67c20c594f3d33bb5422f7"} Mar 20 11:36:05 crc kubenswrapper[4846]: I0320 11:36:05.197312 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42cc748eef09e4d65c9df2e952d44f17290e20653f67c20c594f3d33bb5422f7" Mar 20 11:36:05 crc kubenswrapper[4846]: I0320 11:36:05.197318 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566776-jc9wx" Mar 20 11:36:05 crc kubenswrapper[4846]: I0320 11:36:05.589813 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566770-hq99r"] Mar 20 11:36:05 crc kubenswrapper[4846]: I0320 11:36:05.599264 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566770-hq99r"] Mar 20 11:36:07 crc kubenswrapper[4846]: I0320 11:36:07.331477 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76bc9120-ef21-47ca-ab65-8855bea9dffc" path="/var/lib/kubelet/pods/76bc9120-ef21-47ca-ab65-8855bea9dffc/volumes" Mar 20 11:36:09 crc kubenswrapper[4846]: I0320 11:36:09.678366 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:36:09 crc kubenswrapper[4846]: I0320 11:36:09.678454 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:36:09 crc kubenswrapper[4846]: I0320 11:36:09.678516 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:36:09 crc kubenswrapper[4846]: I0320 11:36:09.679271 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:36:09 crc kubenswrapper[4846]: I0320 11:36:09.679343 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" gracePeriod=600 Mar 20 11:36:09 crc kubenswrapper[4846]: E0320 11:36:09.813283 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:36:10 crc kubenswrapper[4846]: I0320 11:36:10.237279 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" exitCode=0 Mar 20 11:36:10 crc kubenswrapper[4846]: I0320 11:36:10.237359 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72"} Mar 20 11:36:10 crc kubenswrapper[4846]: I0320 11:36:10.237430 4846 scope.go:117] "RemoveContainer" containerID="90a7aa4b08e336909973f1e278c72f0cf1daa26943f9b2d80287000fe2d3a40e" Mar 20 11:36:10 crc kubenswrapper[4846]: I0320 11:36:10.238206 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:36:10 crc kubenswrapper[4846]: E0320 11:36:10.238513 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:36:15 crc kubenswrapper[4846]: I0320 11:36:15.095676 4846 scope.go:117] "RemoveContainer" containerID="1ecb51733fa176a95c9983212d9cac23d49ce05d33b89368fdcc0f9bc7283931" Mar 20 11:36:25 crc kubenswrapper[4846]: I0320 11:36:25.322424 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:36:25 crc kubenswrapper[4846]: E0320 11:36:25.323551 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:36:36 crc kubenswrapper[4846]: I0320 11:36:36.322120 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:36:36 crc kubenswrapper[4846]: E0320 11:36:36.323230 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.059381 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:37 crc kubenswrapper[4846]: E0320 11:36:37.059803 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" containerName="oc" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.059818 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" containerName="oc" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.060020 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" containerName="oc" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.068263 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.074071 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.240523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.240983 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmkwj\" (UniqueName: \"kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.241130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.342858 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmkwj\" (UniqueName: \"kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.342973 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.343038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.343664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.343853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.367469 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmkwj\" (UniqueName: \"kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj\") pod \"community-operators-wd64l\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.402503 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:37 crc kubenswrapper[4846]: I0320 11:36:37.981122 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:38 crc kubenswrapper[4846]: I0320 11:36:38.470929 4846 generic.go:334] "Generic (PLEG): container finished" podID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerID="99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413" exitCode=0 Mar 20 11:36:38 crc kubenswrapper[4846]: I0320 11:36:38.470985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerDied","Data":"99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413"} Mar 20 11:36:38 crc kubenswrapper[4846]: I0320 11:36:38.471031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerStarted","Data":"af59b5865f8968278393b0e579b113f5dd713740d1141e539483183c760cf659"} Mar 20 11:36:39 crc kubenswrapper[4846]: I0320 11:36:39.482243 4846 generic.go:334] "Generic (PLEG): container finished" podID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerID="0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39" exitCode=0 Mar 20 11:36:39 crc kubenswrapper[4846]: I0320 11:36:39.482323 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerDied","Data":"0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39"} Mar 20 11:36:40 crc kubenswrapper[4846]: I0320 11:36:40.493148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerStarted","Data":"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf"} Mar 20 11:36:40 crc kubenswrapper[4846]: I0320 11:36:40.517324 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wd64l" podStartSLOduration=1.90760032 podStartE2EDuration="3.5173028s" podCreationTimestamp="2026-03-20 11:36:37 +0000 UTC" firstStartedPulling="2026-03-20 11:36:38.472945032 +0000 UTC m=+2369.839353268" lastFinishedPulling="2026-03-20 11:36:40.082647512 +0000 UTC m=+2371.449055748" observedRunningTime="2026-03-20 11:36:40.513081357 +0000 UTC m=+2371.879489603" watchObservedRunningTime="2026-03-20 11:36:40.5173028 +0000 UTC m=+2371.883711036" Mar 20 11:36:47 crc kubenswrapper[4846]: I0320 11:36:47.403320 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:47 crc kubenswrapper[4846]: I0320 11:36:47.404367 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:47 crc kubenswrapper[4846]: I0320 11:36:47.455829 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:47 crc kubenswrapper[4846]: I0320 11:36:47.591121 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:47 crc kubenswrapper[4846]: I0320 11:36:47.695774 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:49 crc kubenswrapper[4846]: I0320 11:36:49.328429 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:36:49 crc kubenswrapper[4846]: E0320 11:36:49.328682 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:36:49 crc kubenswrapper[4846]: I0320 11:36:49.565804 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wd64l" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="registry-server" containerID="cri-o://6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf" gracePeriod=2 Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.093695 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.259346 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities\") pod \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.260004 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmkwj\" (UniqueName: \"kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj\") pod \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.260177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content\") pod \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\" (UID: \"14ae7c17-6673-47fc-9386-9343e5a8cfaf\") " Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.260848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities" (OuterVolumeSpecName: "utilities") pod "14ae7c17-6673-47fc-9386-9343e5a8cfaf" (UID: "14ae7c17-6673-47fc-9386-9343e5a8cfaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.267168 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj" (OuterVolumeSpecName: "kube-api-access-mmkwj") pod "14ae7c17-6673-47fc-9386-9343e5a8cfaf" (UID: "14ae7c17-6673-47fc-9386-9343e5a8cfaf"). InnerVolumeSpecName "kube-api-access-mmkwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.362009 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.362056 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmkwj\" (UniqueName: \"kubernetes.io/projected/14ae7c17-6673-47fc-9386-9343e5a8cfaf-kube-api-access-mmkwj\") on node \"crc\" DevicePath \"\"" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.590403 4846 generic.go:334] "Generic (PLEG): container finished" podID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerID="6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf" exitCode=0 Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.590486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerDied","Data":"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf"} Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.590567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd64l" event={"ID":"14ae7c17-6673-47fc-9386-9343e5a8cfaf","Type":"ContainerDied","Data":"af59b5865f8968278393b0e579b113f5dd713740d1141e539483183c760cf659"} Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.590592 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd64l" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.590602 4846 scope.go:117] "RemoveContainer" containerID="6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.623078 4846 scope.go:117] "RemoveContainer" containerID="0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.649426 4846 scope.go:117] "RemoveContainer" containerID="99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.674016 4846 scope.go:117] "RemoveContainer" containerID="6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf" Mar 20 11:36:50 crc kubenswrapper[4846]: E0320 11:36:50.674776 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf\": container with ID starting with 6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf not found: ID does not exist" containerID="6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.674820 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf"} err="failed to get container status \"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf\": rpc error: code = NotFound desc = could not find container \"6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf\": container with ID starting with 6854041bf7189f3edc7bfcd2183f5fdbe6f4f3d2edbfab653511c1c7a199bcdf not found: ID does not exist" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.674852 4846 scope.go:117] "RemoveContainer" containerID="0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39" Mar 20 11:36:50 crc kubenswrapper[4846]: E0320 11:36:50.675443 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39\": container with ID starting with 0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39 not found: ID does not exist" containerID="0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.675510 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39"} err="failed to get container status \"0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39\": rpc error: code = NotFound desc = could not find container \"0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39\": container with ID starting with 0c4ab1587a3a60387d1454f01176b3428bd43dc7eddb3d8b00cedb7a0bfeec39 not found: ID does not exist" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.675547 4846 scope.go:117] "RemoveContainer" containerID="99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413" Mar 20 11:36:50 crc kubenswrapper[4846]: E0320 11:36:50.675987 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413\": container with ID starting with 99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413 not found: ID does not exist" containerID="99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.676052 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413"} err="failed to get container status \"99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413\": rpc error: code = NotFound desc = could not find container \"99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413\": container with ID starting with 99c22b30dc1e8c5bc874b09c74717e0e18a7b291612b1e425679b5b5d669a413 not found: ID does not exist" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.706738 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14ae7c17-6673-47fc-9386-9343e5a8cfaf" (UID: "14ae7c17-6673-47fc-9386-9343e5a8cfaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.767918 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae7c17-6673-47fc-9386-9343e5a8cfaf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.930089 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:50 crc kubenswrapper[4846]: I0320 11:36:50.936752 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wd64l"] Mar 20 11:36:51 crc kubenswrapper[4846]: I0320 11:36:51.333620 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" path="/var/lib/kubelet/pods/14ae7c17-6673-47fc-9386-9343e5a8cfaf/volumes" Mar 20 11:37:01 crc kubenswrapper[4846]: I0320 11:37:01.323179 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:37:01 crc kubenswrapper[4846]: E0320 11:37:01.324299 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:37:14 crc kubenswrapper[4846]: I0320 11:37:14.322264 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:37:14 crc kubenswrapper[4846]: E0320 11:37:14.323152 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:37:27 crc kubenswrapper[4846]: I0320 11:37:27.322200 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:37:27 crc kubenswrapper[4846]: E0320 11:37:27.323109 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:37:39 crc kubenswrapper[4846]: I0320 11:37:39.328009 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:37:39 crc kubenswrapper[4846]: E0320 11:37:39.329365 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:37:53 crc kubenswrapper[4846]: I0320 11:37:53.322501 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:37:53 crc kubenswrapper[4846]: E0320 11:37:53.323467 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.160421 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566778-fww5g"] Mar 20 11:38:00 crc kubenswrapper[4846]: E0320 11:38:00.161662 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="extract-content" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.161678 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="extract-content" Mar 20 11:38:00 crc kubenswrapper[4846]: E0320 11:38:00.161705 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="extract-utilities" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.161712 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="extract-utilities" Mar 20 11:38:00 crc kubenswrapper[4846]: E0320 11:38:00.161731 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="registry-server" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.161737 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="registry-server" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.161919 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae7c17-6673-47fc-9386-9343e5a8cfaf" containerName="registry-server" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.162573 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.165730 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.165802 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.166133 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.169952 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566778-fww5g"] Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.238691 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpksb\" (UniqueName: \"kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb\") pod \"auto-csr-approver-29566778-fww5g\" (UID: \"b080f252-180f-420a-b48f-17e2840a9bc3\") " pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.339981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpksb\" (UniqueName: \"kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb\") pod \"auto-csr-approver-29566778-fww5g\" (UID: \"b080f252-180f-420a-b48f-17e2840a9bc3\") " pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.363656 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpksb\" (UniqueName: \"kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb\") pod \"auto-csr-approver-29566778-fww5g\" (UID: \"b080f252-180f-420a-b48f-17e2840a9bc3\") " pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:00 crc kubenswrapper[4846]: I0320 11:38:00.484659 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:01 crc kubenswrapper[4846]: I0320 11:38:01.340840 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566778-fww5g"] Mar 20 11:38:01 crc kubenswrapper[4846]: I0320 11:38:01.392631 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566778-fww5g" event={"ID":"b080f252-180f-420a-b48f-17e2840a9bc3","Type":"ContainerStarted","Data":"2232d38fae32eb87ab911709fbcaefca5fcac6e7ffc15387cffbbcce58726598"} Mar 20 11:38:03 crc kubenswrapper[4846]: I0320 11:38:03.420000 4846 generic.go:334] "Generic (PLEG): container finished" podID="b080f252-180f-420a-b48f-17e2840a9bc3" containerID="84e7a95cec7a2cd79a1822a58fbc2c735480578a513fee80b2d070251976f899" exitCode=0 Mar 20 11:38:03 crc kubenswrapper[4846]: I0320 11:38:03.420921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566778-fww5g" event={"ID":"b080f252-180f-420a-b48f-17e2840a9bc3","Type":"ContainerDied","Data":"84e7a95cec7a2cd79a1822a58fbc2c735480578a513fee80b2d070251976f899"} Mar 20 11:38:04 crc kubenswrapper[4846]: I0320 11:38:04.723494 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:04 crc kubenswrapper[4846]: I0320 11:38:04.919940 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpksb\" (UniqueName: \"kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb\") pod \"b080f252-180f-420a-b48f-17e2840a9bc3\" (UID: \"b080f252-180f-420a-b48f-17e2840a9bc3\") " Mar 20 11:38:04 crc kubenswrapper[4846]: I0320 11:38:04.927387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb" (OuterVolumeSpecName: "kube-api-access-rpksb") pod "b080f252-180f-420a-b48f-17e2840a9bc3" (UID: "b080f252-180f-420a-b48f-17e2840a9bc3"). InnerVolumeSpecName "kube-api-access-rpksb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.021659 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpksb\" (UniqueName: \"kubernetes.io/projected/b080f252-180f-420a-b48f-17e2840a9bc3-kube-api-access-rpksb\") on node \"crc\" DevicePath \"\"" Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.439783 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566778-fww5g" event={"ID":"b080f252-180f-420a-b48f-17e2840a9bc3","Type":"ContainerDied","Data":"2232d38fae32eb87ab911709fbcaefca5fcac6e7ffc15387cffbbcce58726598"} Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.439849 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2232d38fae32eb87ab911709fbcaefca5fcac6e7ffc15387cffbbcce58726598" Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.439888 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566778-fww5g" Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.798318 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566772-2zq6x"] Mar 20 11:38:05 crc kubenswrapper[4846]: I0320 11:38:05.803591 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566772-2zq6x"] Mar 20 11:38:07 crc kubenswrapper[4846]: I0320 11:38:07.322634 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:38:07 crc kubenswrapper[4846]: E0320 11:38:07.323398 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:38:07 crc kubenswrapper[4846]: I0320 11:38:07.334278 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de333c6-d33e-41e2-bf7e-2c9a137e37c3" path="/var/lib/kubelet/pods/1de333c6-d33e-41e2-bf7e-2c9a137e37c3/volumes" Mar 20 11:38:15 crc kubenswrapper[4846]: I0320 11:38:15.207469 4846 scope.go:117] "RemoveContainer" containerID="7ee2fc2b5ed323fcb85163f190103adc7135fbcf245bb405a07efa7bf509335d" Mar 20 11:38:22 crc kubenswrapper[4846]: I0320 11:38:22.323137 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:38:22 crc kubenswrapper[4846]: E0320 11:38:22.324063 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:38:34 crc kubenswrapper[4846]: I0320 11:38:34.322728 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:38:34 crc kubenswrapper[4846]: E0320 11:38:34.324333 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:38:48 crc kubenswrapper[4846]: I0320 11:38:48.323027 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:38:48 crc kubenswrapper[4846]: E0320 11:38:48.325623 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:39:03 crc kubenswrapper[4846]: I0320 11:39:03.322427 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:39:03 crc kubenswrapper[4846]: E0320 11:39:03.323434 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.435725 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:06 crc kubenswrapper[4846]: E0320 11:39:06.436694 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b080f252-180f-420a-b48f-17e2840a9bc3" containerName="oc" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.436714 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b080f252-180f-420a-b48f-17e2840a9bc3" containerName="oc" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.436929 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b080f252-180f-420a-b48f-17e2840a9bc3" containerName="oc" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.438224 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.448653 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.601143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.601217 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwjmh\" (UniqueName: \"kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.601261 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.703258 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwjmh\" (UniqueName: \"kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.703358 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.703448 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.703964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.704020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.732808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwjmh\" (UniqueName: \"kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh\") pod \"certified-operators-dhgmr\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:06 crc kubenswrapper[4846]: I0320 11:39:06.761623 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:07 crc kubenswrapper[4846]: I0320 11:39:07.269972 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:07 crc kubenswrapper[4846]: I0320 11:39:07.959471 4846 generic.go:334] "Generic (PLEG): container finished" podID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerID="3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad" exitCode=0 Mar 20 11:39:07 crc kubenswrapper[4846]: I0320 11:39:07.959602 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerDied","Data":"3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad"} Mar 20 11:39:07 crc kubenswrapper[4846]: I0320 11:39:07.959860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerStarted","Data":"fe0ade37e501d57d44c0f43e833d37dba2569f36bad6b206c75019cecffcffee"} Mar 20 11:39:07 crc kubenswrapper[4846]: I0320 11:39:07.961722 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:39:08 crc kubenswrapper[4846]: I0320 11:39:08.970830 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerStarted","Data":"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2"} Mar 20 11:39:09 crc kubenswrapper[4846]: I0320 11:39:09.981837 4846 generic.go:334] "Generic (PLEG): container finished" podID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerID="8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2" exitCode=0 Mar 20 11:39:09 crc kubenswrapper[4846]: I0320 11:39:09.981947 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerDied","Data":"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2"} Mar 20 11:39:10 crc kubenswrapper[4846]: I0320 11:39:10.993937 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerStarted","Data":"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b"} Mar 20 11:39:11 crc kubenswrapper[4846]: I0320 11:39:11.018242 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dhgmr" podStartSLOduration=2.622431448 podStartE2EDuration="5.018222316s" podCreationTimestamp="2026-03-20 11:39:06 +0000 UTC" firstStartedPulling="2026-03-20 11:39:07.961399241 +0000 UTC m=+2519.327807477" lastFinishedPulling="2026-03-20 11:39:10.357190109 +0000 UTC m=+2521.723598345" observedRunningTime="2026-03-20 11:39:11.012800053 +0000 UTC m=+2522.379208289" watchObservedRunningTime="2026-03-20 11:39:11.018222316 +0000 UTC m=+2522.384630552" Mar 20 11:39:16 crc kubenswrapper[4846]: I0320 11:39:16.323191 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:39:16 crc kubenswrapper[4846]: E0320 11:39:16.324289 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:39:16 crc kubenswrapper[4846]: I0320 11:39:16.761888 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:16 crc kubenswrapper[4846]: I0320 11:39:16.762023 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:16 crc kubenswrapper[4846]: I0320 11:39:16.817726 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:17 crc kubenswrapper[4846]: I0320 11:39:17.103005 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:17 crc kubenswrapper[4846]: I0320 11:39:17.159605 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.073504 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dhgmr" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="registry-server" containerID="cri-o://5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b" gracePeriod=2 Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.826984 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.935622 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content\") pod \"cb1363bd-d57c-4129-955a-a74abf620aa3\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.935683 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwjmh\" (UniqueName: \"kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh\") pod \"cb1363bd-d57c-4129-955a-a74abf620aa3\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.935707 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities\") pod \"cb1363bd-d57c-4129-955a-a74abf620aa3\" (UID: \"cb1363bd-d57c-4129-955a-a74abf620aa3\") " Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.936743 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities" (OuterVolumeSpecName: "utilities") pod "cb1363bd-d57c-4129-955a-a74abf620aa3" (UID: "cb1363bd-d57c-4129-955a-a74abf620aa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.949194 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh" (OuterVolumeSpecName: "kube-api-access-dwjmh") pod "cb1363bd-d57c-4129-955a-a74abf620aa3" (UID: "cb1363bd-d57c-4129-955a-a74abf620aa3"). InnerVolumeSpecName "kube-api-access-dwjmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:39:19 crc kubenswrapper[4846]: I0320 11:39:19.998783 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb1363bd-d57c-4129-955a-a74abf620aa3" (UID: "cb1363bd-d57c-4129-955a-a74abf620aa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.038289 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.038355 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1363bd-d57c-4129-955a-a74abf620aa3-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.038370 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwjmh\" (UniqueName: \"kubernetes.io/projected/cb1363bd-d57c-4129-955a-a74abf620aa3-kube-api-access-dwjmh\") on node \"crc\" DevicePath \"\"" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.082746 4846 generic.go:334] "Generic (PLEG): container finished" podID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerID="5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b" exitCode=0 Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.082823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerDied","Data":"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b"} Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.082903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dhgmr" event={"ID":"cb1363bd-d57c-4129-955a-a74abf620aa3","Type":"ContainerDied","Data":"fe0ade37e501d57d44c0f43e833d37dba2569f36bad6b206c75019cecffcffee"} Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.082940 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dhgmr" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.082947 4846 scope.go:117] "RemoveContainer" containerID="5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.111113 4846 scope.go:117] "RemoveContainer" containerID="8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.129674 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.138695 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dhgmr"] Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.146241 4846 scope.go:117] "RemoveContainer" containerID="3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.170051 4846 scope.go:117] "RemoveContainer" containerID="5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b" Mar 20 11:39:20 crc kubenswrapper[4846]: E0320 11:39:20.170565 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b\": container with ID starting with 5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b not found: ID does not exist" containerID="5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.170600 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b"} err="failed to get container status \"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b\": rpc error: code = NotFound desc = could not find container \"5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b\": container with ID starting with 5846200084fd6cc497ab41a85e2aee138cb91708a8862d75f7a0d1d5fe71de9b not found: ID does not exist" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.170629 4846 scope.go:117] "RemoveContainer" containerID="8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2" Mar 20 11:39:20 crc kubenswrapper[4846]: E0320 11:39:20.171218 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2\": container with ID starting with 8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2 not found: ID does not exist" containerID="8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.171267 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2"} err="failed to get container status \"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2\": rpc error: code = NotFound desc = could not find container \"8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2\": container with ID starting with 8568547644edb3c6793f8c12adb3c8650f1b1b326e010f8022811659c118ebc2 not found: ID does not exist" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.171317 4846 scope.go:117] "RemoveContainer" containerID="3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad" Mar 20 11:39:20 crc kubenswrapper[4846]: E0320 11:39:20.171657 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad\": container with ID starting with 3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad not found: ID does not exist" containerID="3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad" Mar 20 11:39:20 crc kubenswrapper[4846]: I0320 11:39:20.171691 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad"} err="failed to get container status \"3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad\": rpc error: code = NotFound desc = could not find container \"3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad\": container with ID starting with 3a5fd9d9259054698b02072b9a8ecffb7041833fdbc095a098ac3731624bbaad not found: ID does not exist" Mar 20 11:39:21 crc kubenswrapper[4846]: I0320 11:39:21.339566 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" path="/var/lib/kubelet/pods/cb1363bd-d57c-4129-955a-a74abf620aa3/volumes" Mar 20 11:39:29 crc kubenswrapper[4846]: I0320 11:39:29.323230 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:39:29 crc kubenswrapper[4846]: E0320 11:39:29.324598 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:39:41 crc kubenswrapper[4846]: I0320 11:39:41.323639 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:39:41 crc kubenswrapper[4846]: E0320 11:39:41.324843 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:39:55 crc kubenswrapper[4846]: I0320 11:39:55.323242 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:39:55 crc kubenswrapper[4846]: E0320 11:39:55.324182 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.153060 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566780-gxgf4"] Mar 20 11:40:00 crc kubenswrapper[4846]: E0320 11:40:00.153889 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="extract-content" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.154017 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="extract-content" Mar 20 11:40:00 crc kubenswrapper[4846]: E0320 11:40:00.154051 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="registry-server" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.154060 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="registry-server" Mar 20 11:40:00 crc kubenswrapper[4846]: E0320 11:40:00.154073 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="extract-utilities" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.154085 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="extract-utilities" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.154270 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb1363bd-d57c-4129-955a-a74abf620aa3" containerName="registry-server" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.154885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.157272 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.157956 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.158658 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.172188 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566780-gxgf4"] Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.205941 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kstcn\" (UniqueName: \"kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn\") pod \"auto-csr-approver-29566780-gxgf4\" (UID: \"63aac89a-b859-4734-8c82-0fd5a7bd373a\") " pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.307581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kstcn\" (UniqueName: \"kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn\") pod \"auto-csr-approver-29566780-gxgf4\" (UID: \"63aac89a-b859-4734-8c82-0fd5a7bd373a\") " pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.328760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kstcn\" (UniqueName: \"kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn\") pod \"auto-csr-approver-29566780-gxgf4\" (UID: \"63aac89a-b859-4734-8c82-0fd5a7bd373a\") " pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.479163 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:00 crc kubenswrapper[4846]: I0320 11:40:00.942868 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566780-gxgf4"] Mar 20 11:40:01 crc kubenswrapper[4846]: I0320 11:40:01.407308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" event={"ID":"63aac89a-b859-4734-8c82-0fd5a7bd373a","Type":"ContainerStarted","Data":"2b47dbf8dd88f262d4916e046ac790a2a038fa3ce8db9efa508fb35d8844fac4"} Mar 20 11:40:03 crc kubenswrapper[4846]: I0320 11:40:03.426008 4846 generic.go:334] "Generic (PLEG): container finished" podID="63aac89a-b859-4734-8c82-0fd5a7bd373a" containerID="a22a3f36a3684fe48674376a80e47d3ae371dd01c3b0392123b9dfcbc296d662" exitCode=0 Mar 20 11:40:03 crc kubenswrapper[4846]: I0320 11:40:03.426074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" event={"ID":"63aac89a-b859-4734-8c82-0fd5a7bd373a","Type":"ContainerDied","Data":"a22a3f36a3684fe48674376a80e47d3ae371dd01c3b0392123b9dfcbc296d662"} Mar 20 11:40:04 crc kubenswrapper[4846]: I0320 11:40:04.724973 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:04 crc kubenswrapper[4846]: I0320 11:40:04.806354 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kstcn\" (UniqueName: \"kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn\") pod \"63aac89a-b859-4734-8c82-0fd5a7bd373a\" (UID: \"63aac89a-b859-4734-8c82-0fd5a7bd373a\") " Mar 20 11:40:04 crc kubenswrapper[4846]: I0320 11:40:04.813869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn" (OuterVolumeSpecName: "kube-api-access-kstcn") pod "63aac89a-b859-4734-8c82-0fd5a7bd373a" (UID: "63aac89a-b859-4734-8c82-0fd5a7bd373a"). InnerVolumeSpecName "kube-api-access-kstcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:40:04 crc kubenswrapper[4846]: I0320 11:40:04.908777 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kstcn\" (UniqueName: \"kubernetes.io/projected/63aac89a-b859-4734-8c82-0fd5a7bd373a-kube-api-access-kstcn\") on node \"crc\" DevicePath \"\"" Mar 20 11:40:05 crc kubenswrapper[4846]: I0320 11:40:05.443593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" event={"ID":"63aac89a-b859-4734-8c82-0fd5a7bd373a","Type":"ContainerDied","Data":"2b47dbf8dd88f262d4916e046ac790a2a038fa3ce8db9efa508fb35d8844fac4"} Mar 20 11:40:05 crc kubenswrapper[4846]: I0320 11:40:05.443644 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b47dbf8dd88f262d4916e046ac790a2a038fa3ce8db9efa508fb35d8844fac4" Mar 20 11:40:05 crc kubenswrapper[4846]: I0320 11:40:05.443675 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566780-gxgf4" Mar 20 11:40:05 crc kubenswrapper[4846]: I0320 11:40:05.801829 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566774-thkdc"] Mar 20 11:40:05 crc kubenswrapper[4846]: I0320 11:40:05.807796 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566774-thkdc"] Mar 20 11:40:07 crc kubenswrapper[4846]: I0320 11:40:07.332166 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d147a127-88f4-42bb-9879-5f2ff1ace4eb" path="/var/lib/kubelet/pods/d147a127-88f4-42bb-9879-5f2ff1ace4eb/volumes" Mar 20 11:40:10 crc kubenswrapper[4846]: I0320 11:40:10.323596 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:40:10 crc kubenswrapper[4846]: E0320 11:40:10.324464 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:40:15 crc kubenswrapper[4846]: I0320 11:40:15.296887 4846 scope.go:117] "RemoveContainer" containerID="f40820e0af62d8586713fc6eb7bff38f6a48bec8a116c797b2099dfb160a3ab5" Mar 20 11:40:25 crc kubenswrapper[4846]: I0320 11:40:25.326239 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:40:25 crc kubenswrapper[4846]: E0320 11:40:25.327714 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:40:40 crc kubenswrapper[4846]: I0320 11:40:40.323278 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:40:40 crc kubenswrapper[4846]: E0320 11:40:40.324645 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:40:51 crc kubenswrapper[4846]: I0320 11:40:51.327400 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:40:51 crc kubenswrapper[4846]: E0320 11:40:51.328630 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:41:04 crc kubenswrapper[4846]: I0320 11:41:04.323499 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:41:04 crc kubenswrapper[4846]: E0320 11:41:04.324492 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:41:18 crc kubenswrapper[4846]: I0320 11:41:18.322417 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:41:18 crc kubenswrapper[4846]: I0320 11:41:18.978709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4"} Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.145088 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566782-glcd2"] Mar 20 11:42:00 crc kubenswrapper[4846]: E0320 11:42:00.146204 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63aac89a-b859-4734-8c82-0fd5a7bd373a" containerName="oc" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.146220 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="63aac89a-b859-4734-8c82-0fd5a7bd373a" containerName="oc" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.146354 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="63aac89a-b859-4734-8c82-0fd5a7bd373a" containerName="oc" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.147008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.149937 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.150194 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.159239 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566782-glcd2"] Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.161334 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.269469 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmbbb\" (UniqueName: \"kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb\") pod \"auto-csr-approver-29566782-glcd2\" (UID: \"d29560ab-925d-4746-90ea-fb3f295e90c0\") " pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.371376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmbbb\" (UniqueName: \"kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb\") pod \"auto-csr-approver-29566782-glcd2\" (UID: \"d29560ab-925d-4746-90ea-fb3f295e90c0\") " pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.393330 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmbbb\" (UniqueName: \"kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb\") pod \"auto-csr-approver-29566782-glcd2\" (UID: \"d29560ab-925d-4746-90ea-fb3f295e90c0\") " pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.473408 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:00 crc kubenswrapper[4846]: I0320 11:42:00.890091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566782-glcd2"] Mar 20 11:42:01 crc kubenswrapper[4846]: I0320 11:42:01.316626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566782-glcd2" event={"ID":"d29560ab-925d-4746-90ea-fb3f295e90c0","Type":"ContainerStarted","Data":"ab9648df7239140c48789f1d5cd991dce4ff6251f84167bb6919a14353f18271"} Mar 20 11:42:02 crc kubenswrapper[4846]: I0320 11:42:02.327589 4846 generic.go:334] "Generic (PLEG): container finished" podID="d29560ab-925d-4746-90ea-fb3f295e90c0" containerID="92c11d91c7ed5cfa83ec55250b78a8d26e87b63a161a597d4942ce5b78d99c42" exitCode=0 Mar 20 11:42:02 crc kubenswrapper[4846]: I0320 11:42:02.327715 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566782-glcd2" event={"ID":"d29560ab-925d-4746-90ea-fb3f295e90c0","Type":"ContainerDied","Data":"92c11d91c7ed5cfa83ec55250b78a8d26e87b63a161a597d4942ce5b78d99c42"} Mar 20 11:42:03 crc kubenswrapper[4846]: I0320 11:42:03.632096 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:03 crc kubenswrapper[4846]: I0320 11:42:03.724156 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmbbb\" (UniqueName: \"kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb\") pod \"d29560ab-925d-4746-90ea-fb3f295e90c0\" (UID: \"d29560ab-925d-4746-90ea-fb3f295e90c0\") " Mar 20 11:42:03 crc kubenswrapper[4846]: I0320 11:42:03.732127 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb" (OuterVolumeSpecName: "kube-api-access-pmbbb") pod "d29560ab-925d-4746-90ea-fb3f295e90c0" (UID: "d29560ab-925d-4746-90ea-fb3f295e90c0"). InnerVolumeSpecName "kube-api-access-pmbbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:42:03 crc kubenswrapper[4846]: I0320 11:42:03.826239 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmbbb\" (UniqueName: \"kubernetes.io/projected/d29560ab-925d-4746-90ea-fb3f295e90c0-kube-api-access-pmbbb\") on node \"crc\" DevicePath \"\"" Mar 20 11:42:04 crc kubenswrapper[4846]: I0320 11:42:04.346882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566782-glcd2" event={"ID":"d29560ab-925d-4746-90ea-fb3f295e90c0","Type":"ContainerDied","Data":"ab9648df7239140c48789f1d5cd991dce4ff6251f84167bb6919a14353f18271"} Mar 20 11:42:04 crc kubenswrapper[4846]: I0320 11:42:04.347399 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab9648df7239140c48789f1d5cd991dce4ff6251f84167bb6919a14353f18271" Mar 20 11:42:04 crc kubenswrapper[4846]: I0320 11:42:04.346993 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566782-glcd2" Mar 20 11:42:04 crc kubenswrapper[4846]: I0320 11:42:04.714352 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566776-jc9wx"] Mar 20 11:42:04 crc kubenswrapper[4846]: I0320 11:42:04.721789 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566776-jc9wx"] Mar 20 11:42:05 crc kubenswrapper[4846]: I0320 11:42:05.335677 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6" path="/var/lib/kubelet/pods/8c36dfd1-4775-4801-b2eb-f1a4d32e9ac6/volumes" Mar 20 11:42:15 crc kubenswrapper[4846]: I0320 11:42:15.388769 4846 scope.go:117] "RemoveContainer" containerID="8b886ad83c6a569bc39cb8dd3efb39bdac5c959c39b0d126de1699cf04253bcb" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.698528 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:42:58 crc kubenswrapper[4846]: E0320 11:42:58.702196 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29560ab-925d-4746-90ea-fb3f295e90c0" containerName="oc" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.702219 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29560ab-925d-4746-90ea-fb3f295e90c0" containerName="oc" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.702408 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29560ab-925d-4746-90ea-fb3f295e90c0" containerName="oc" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.703568 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.715154 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.818008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.818075 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvx5\" (UniqueName: \"kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.818105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.919333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.919383 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvx5\" (UniqueName: \"kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.919411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.919984 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.920026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:58 crc kubenswrapper[4846]: I0320 11:42:58.943477 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvx5\" (UniqueName: \"kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5\") pod \"redhat-marketplace-vdvmq\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:59 crc kubenswrapper[4846]: I0320 11:42:59.035735 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:42:59 crc kubenswrapper[4846]: I0320 11:42:59.513242 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:42:59 crc kubenswrapper[4846]: I0320 11:42:59.802127 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerID="a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7" exitCode=0 Mar 20 11:42:59 crc kubenswrapper[4846]: I0320 11:42:59.802200 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerDied","Data":"a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7"} Mar 20 11:42:59 crc kubenswrapper[4846]: I0320 11:42:59.802246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerStarted","Data":"94dcd34d7c1abfca284d3cb687be828c0eb38fdd992ddf00ae9b8dfa36307bf1"} Mar 20 11:43:01 crc kubenswrapper[4846]: I0320 11:43:01.820579 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerID="6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04" exitCode=0 Mar 20 11:43:01 crc kubenswrapper[4846]: I0320 11:43:01.820674 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerDied","Data":"6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04"} Mar 20 11:43:02 crc kubenswrapper[4846]: I0320 11:43:02.832377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerStarted","Data":"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5"} Mar 20 11:43:02 crc kubenswrapper[4846]: I0320 11:43:02.862250 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vdvmq" podStartSLOduration=2.310701919 podStartE2EDuration="4.862224728s" podCreationTimestamp="2026-03-20 11:42:58 +0000 UTC" firstStartedPulling="2026-03-20 11:42:59.804621431 +0000 UTC m=+2751.171029657" lastFinishedPulling="2026-03-20 11:43:02.35614423 +0000 UTC m=+2753.722552466" observedRunningTime="2026-03-20 11:43:02.853748209 +0000 UTC m=+2754.220156455" watchObservedRunningTime="2026-03-20 11:43:02.862224728 +0000 UTC m=+2754.228632954" Mar 20 11:43:09 crc kubenswrapper[4846]: I0320 11:43:09.036814 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:09 crc kubenswrapper[4846]: I0320 11:43:09.037357 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:09 crc kubenswrapper[4846]: I0320 11:43:09.085185 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:09 crc kubenswrapper[4846]: I0320 11:43:09.939628 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:09 crc kubenswrapper[4846]: I0320 11:43:09.994380 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:43:11 crc kubenswrapper[4846]: I0320 11:43:11.910598 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vdvmq" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="registry-server" containerID="cri-o://91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5" gracePeriod=2 Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.325061 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.454154 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities\") pod \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.454263 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content\") pod \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.454287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvx5\" (UniqueName: \"kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5\") pod \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\" (UID: \"e4d3566f-fe3f-4777-a595-7b7996fb2a8c\") " Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.455268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities" (OuterVolumeSpecName: "utilities") pod "e4d3566f-fe3f-4777-a595-7b7996fb2a8c" (UID: "e4d3566f-fe3f-4777-a595-7b7996fb2a8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.455848 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.461833 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5" (OuterVolumeSpecName: "kube-api-access-qsvx5") pod "e4d3566f-fe3f-4777-a595-7b7996fb2a8c" (UID: "e4d3566f-fe3f-4777-a595-7b7996fb2a8c"). InnerVolumeSpecName "kube-api-access-qsvx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.479310 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4d3566f-fe3f-4777-a595-7b7996fb2a8c" (UID: "e4d3566f-fe3f-4777-a595-7b7996fb2a8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.556994 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.557036 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvx5\" (UniqueName: \"kubernetes.io/projected/e4d3566f-fe3f-4777-a595-7b7996fb2a8c-kube-api-access-qsvx5\") on node \"crc\" DevicePath \"\"" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.940808 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerID="91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5" exitCode=0 Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.940947 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerDied","Data":"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5"} Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.941010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vdvmq" event={"ID":"e4d3566f-fe3f-4777-a595-7b7996fb2a8c","Type":"ContainerDied","Data":"94dcd34d7c1abfca284d3cb687be828c0eb38fdd992ddf00ae9b8dfa36307bf1"} Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.941049 4846 scope.go:117] "RemoveContainer" containerID="91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.941412 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vdvmq" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.979653 4846 scope.go:117] "RemoveContainer" containerID="6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04" Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.987880 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:43:12 crc kubenswrapper[4846]: I0320 11:43:12.996526 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vdvmq"] Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.016263 4846 scope.go:117] "RemoveContainer" containerID="a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.047789 4846 scope.go:117] "RemoveContainer" containerID="91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5" Mar 20 11:43:13 crc kubenswrapper[4846]: E0320 11:43:13.049119 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5\": container with ID starting with 91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5 not found: ID does not exist" containerID="91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.049175 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5"} err="failed to get container status \"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5\": rpc error: code = NotFound desc = could not find container \"91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5\": container with ID starting with 91c9adbee4bafc1e713368ecee0a61c225cbb2a19a00f7ca230c55c7ad603bf5 not found: ID does not exist" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.049209 4846 scope.go:117] "RemoveContainer" containerID="6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04" Mar 20 11:43:13 crc kubenswrapper[4846]: E0320 11:43:13.049810 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04\": container with ID starting with 6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04 not found: ID does not exist" containerID="6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.049835 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04"} err="failed to get container status \"6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04\": rpc error: code = NotFound desc = could not find container \"6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04\": container with ID starting with 6edf2dfa2ff8340949b2fbdc3371d93a4573a8af4129fd20ca223a6781a29b04 not found: ID does not exist" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.049853 4846 scope.go:117] "RemoveContainer" containerID="a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7" Mar 20 11:43:13 crc kubenswrapper[4846]: E0320 11:43:13.050194 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7\": container with ID starting with a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7 not found: ID does not exist" containerID="a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.050249 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7"} err="failed to get container status \"a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7\": rpc error: code = NotFound desc = could not find container \"a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7\": container with ID starting with a880302b9dfe9f08159911a581d9c5fc201cccc4d296f7f01ef30c204c6d29c7 not found: ID does not exist" Mar 20 11:43:13 crc kubenswrapper[4846]: I0320 11:43:13.340087 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" path="/var/lib/kubelet/pods/e4d3566f-fe3f-4777-a595-7b7996fb2a8c/volumes" Mar 20 11:43:39 crc kubenswrapper[4846]: I0320 11:43:39.677252 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:43:39 crc kubenswrapper[4846]: I0320 11:43:39.677987 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.143731 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566784-5mbbf"] Mar 20 11:44:00 crc kubenswrapper[4846]: E0320 11:44:00.144736 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="extract-content" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.144753 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="extract-content" Mar 20 11:44:00 crc kubenswrapper[4846]: E0320 11:44:00.144781 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="extract-utilities" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.144792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="extract-utilities" Mar 20 11:44:00 crc kubenswrapper[4846]: E0320 11:44:00.144817 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="registry-server" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.144825 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="registry-server" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.147259 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4d3566f-fe3f-4777-a595-7b7996fb2a8c" containerName="registry-server" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.148076 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.154780 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.155136 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.155644 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.160715 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566784-5mbbf"] Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.268368 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jjgk\" (UniqueName: \"kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk\") pod \"auto-csr-approver-29566784-5mbbf\" (UID: \"a755a71b-dc8c-4bcd-9144-60454c21d8a4\") " pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.369756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jjgk\" (UniqueName: \"kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk\") pod \"auto-csr-approver-29566784-5mbbf\" (UID: \"a755a71b-dc8c-4bcd-9144-60454c21d8a4\") " pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.389949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jjgk\" (UniqueName: \"kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk\") pod \"auto-csr-approver-29566784-5mbbf\" (UID: \"a755a71b-dc8c-4bcd-9144-60454c21d8a4\") " pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.477248 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:00 crc kubenswrapper[4846]: I0320 11:44:00.897754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566784-5mbbf"] Mar 20 11:44:01 crc kubenswrapper[4846]: I0320 11:44:01.314018 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" event={"ID":"a755a71b-dc8c-4bcd-9144-60454c21d8a4","Type":"ContainerStarted","Data":"b11f5d91a5d39d0db7d518cc2e15be1a21d77ef4ca784d9fadeab8c266ce00e9"} Mar 20 11:44:02 crc kubenswrapper[4846]: I0320 11:44:02.327670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" event={"ID":"a755a71b-dc8c-4bcd-9144-60454c21d8a4","Type":"ContainerStarted","Data":"99b7218166dff07b7fb42f8f4d469029c709c5de20e38764d68a67e5bf62b50e"} Mar 20 11:44:02 crc kubenswrapper[4846]: I0320 11:44:02.346294 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" podStartSLOduration=1.377551264 podStartE2EDuration="2.346274769s" podCreationTimestamp="2026-03-20 11:44:00 +0000 UTC" firstStartedPulling="2026-03-20 11:44:00.910808498 +0000 UTC m=+2812.277216734" lastFinishedPulling="2026-03-20 11:44:01.879532003 +0000 UTC m=+2813.245940239" observedRunningTime="2026-03-20 11:44:02.345253134 +0000 UTC m=+2813.711661360" watchObservedRunningTime="2026-03-20 11:44:02.346274769 +0000 UTC m=+2813.712683005" Mar 20 11:44:03 crc kubenswrapper[4846]: I0320 11:44:03.345997 4846 generic.go:334] "Generic (PLEG): container finished" podID="a755a71b-dc8c-4bcd-9144-60454c21d8a4" containerID="99b7218166dff07b7fb42f8f4d469029c709c5de20e38764d68a67e5bf62b50e" exitCode=0 Mar 20 11:44:03 crc kubenswrapper[4846]: I0320 11:44:03.346146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" event={"ID":"a755a71b-dc8c-4bcd-9144-60454c21d8a4","Type":"ContainerDied","Data":"99b7218166dff07b7fb42f8f4d469029c709c5de20e38764d68a67e5bf62b50e"} Mar 20 11:44:04 crc kubenswrapper[4846]: I0320 11:44:04.681819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:04 crc kubenswrapper[4846]: I0320 11:44:04.739468 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jjgk\" (UniqueName: \"kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk\") pod \"a755a71b-dc8c-4bcd-9144-60454c21d8a4\" (UID: \"a755a71b-dc8c-4bcd-9144-60454c21d8a4\") " Mar 20 11:44:04 crc kubenswrapper[4846]: I0320 11:44:04.747721 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk" (OuterVolumeSpecName: "kube-api-access-4jjgk") pod "a755a71b-dc8c-4bcd-9144-60454c21d8a4" (UID: "a755a71b-dc8c-4bcd-9144-60454c21d8a4"). InnerVolumeSpecName "kube-api-access-4jjgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:44:04 crc kubenswrapper[4846]: I0320 11:44:04.841471 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jjgk\" (UniqueName: \"kubernetes.io/projected/a755a71b-dc8c-4bcd-9144-60454c21d8a4-kube-api-access-4jjgk\") on node \"crc\" DevicePath \"\"" Mar 20 11:44:05 crc kubenswrapper[4846]: I0320 11:44:05.362230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" event={"ID":"a755a71b-dc8c-4bcd-9144-60454c21d8a4","Type":"ContainerDied","Data":"b11f5d91a5d39d0db7d518cc2e15be1a21d77ef4ca784d9fadeab8c266ce00e9"} Mar 20 11:44:05 crc kubenswrapper[4846]: I0320 11:44:05.362283 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b11f5d91a5d39d0db7d518cc2e15be1a21d77ef4ca784d9fadeab8c266ce00e9" Mar 20 11:44:05 crc kubenswrapper[4846]: I0320 11:44:05.362335 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566784-5mbbf" Mar 20 11:44:05 crc kubenswrapper[4846]: I0320 11:44:05.413128 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566778-fww5g"] Mar 20 11:44:05 crc kubenswrapper[4846]: I0320 11:44:05.419711 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566778-fww5g"] Mar 20 11:44:07 crc kubenswrapper[4846]: I0320 11:44:07.333341 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b080f252-180f-420a-b48f-17e2840a9bc3" path="/var/lib/kubelet/pods/b080f252-180f-420a-b48f-17e2840a9bc3/volumes" Mar 20 11:44:09 crc kubenswrapper[4846]: I0320 11:44:09.677808 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:44:09 crc kubenswrapper[4846]: I0320 11:44:09.678402 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:44:15 crc kubenswrapper[4846]: I0320 11:44:15.491762 4846 scope.go:117] "RemoveContainer" containerID="84e7a95cec7a2cd79a1822a58fbc2c735480578a513fee80b2d070251976f899" Mar 20 11:44:39 crc kubenswrapper[4846]: I0320 11:44:39.678326 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:44:39 crc kubenswrapper[4846]: I0320 11:44:39.679180 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:44:39 crc kubenswrapper[4846]: I0320 11:44:39.679245 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:44:39 crc kubenswrapper[4846]: I0320 11:44:39.679980 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:44:39 crc kubenswrapper[4846]: I0320 11:44:39.680055 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4" gracePeriod=600 Mar 20 11:44:40 crc kubenswrapper[4846]: I0320 11:44:40.628524 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4" exitCode=0 Mar 20 11:44:40 crc kubenswrapper[4846]: I0320 11:44:40.628800 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4"} Mar 20 11:44:40 crc kubenswrapper[4846]: I0320 11:44:40.629386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60"} Mar 20 11:44:40 crc kubenswrapper[4846]: I0320 11:44:40.629418 4846 scope.go:117] "RemoveContainer" containerID="65d3286a7d82960c242aa407a003f5b9745cc2395a99a9c24a1e335673950c72" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.147392 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd"] Mar 20 11:45:00 crc kubenswrapper[4846]: E0320 11:45:00.148758 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a755a71b-dc8c-4bcd-9144-60454c21d8a4" containerName="oc" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.148807 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a755a71b-dc8c-4bcd-9144-60454c21d8a4" containerName="oc" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.148990 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a755a71b-dc8c-4bcd-9144-60454c21d8a4" containerName="oc" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.149678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.152412 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.153023 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.157859 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd"] Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.344949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.346031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.346229 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvfqj\" (UniqueName: \"kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.448332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.448991 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.449171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvfqj\" (UniqueName: \"kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.449782 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.456684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.469279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvfqj\" (UniqueName: \"kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj\") pod \"collect-profiles-29566785-nmqwd\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.494436 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:00 crc kubenswrapper[4846]: I0320 11:45:00.930406 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd"] Mar 20 11:45:00 crc kubenswrapper[4846]: W0320 11:45:00.944259 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8768aa20_2e5f_4a04_a79d_81deb3f9dc3e.slice/crio-df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b WatchSource:0}: Error finding container df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b: Status 404 returned error can't find the container with id df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b Mar 20 11:45:01 crc kubenswrapper[4846]: I0320 11:45:01.795496 4846 generic.go:334] "Generic (PLEG): container finished" podID="8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" containerID="b7665f864ee627806a2ec3d1135d64e04fcb34a113cc56f74ecfe64438718b7c" exitCode=0 Mar 20 11:45:01 crc kubenswrapper[4846]: I0320 11:45:01.796070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" event={"ID":"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e","Type":"ContainerDied","Data":"b7665f864ee627806a2ec3d1135d64e04fcb34a113cc56f74ecfe64438718b7c"} Mar 20 11:45:01 crc kubenswrapper[4846]: I0320 11:45:01.796106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" event={"ID":"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e","Type":"ContainerStarted","Data":"df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b"} Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.132159 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.231878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvfqj\" (UniqueName: \"kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj\") pod \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.232503 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume\") pod \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.232597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume\") pod \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\" (UID: \"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e\") " Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.233707 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume" (OuterVolumeSpecName: "config-volume") pod "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" (UID: "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.240514 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj" (OuterVolumeSpecName: "kube-api-access-gvfqj") pod "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" (UID: "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e"). InnerVolumeSpecName "kube-api-access-gvfqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.241008 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" (UID: "8768aa20-2e5f-4a04-a79d-81deb3f9dc3e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.334817 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvfqj\" (UniqueName: \"kubernetes.io/projected/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-kube-api-access-gvfqj\") on node \"crc\" DevicePath \"\"" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.334864 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.334874 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8768aa20-2e5f-4a04-a79d-81deb3f9dc3e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.825158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" event={"ID":"8768aa20-2e5f-4a04-a79d-81deb3f9dc3e","Type":"ContainerDied","Data":"df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b"} Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.825229 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df235f4a0aa404af13d2db7ad4394176d81562ab7bde1ad2fd073a328d271e9b" Mar 20 11:45:03 crc kubenswrapper[4846]: I0320 11:45:03.825281 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566785-nmqwd" Mar 20 11:45:04 crc kubenswrapper[4846]: I0320 11:45:04.222643 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z"] Mar 20 11:45:04 crc kubenswrapper[4846]: I0320 11:45:04.228472 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566740-85c7z"] Mar 20 11:45:05 crc kubenswrapper[4846]: I0320 11:45:05.333586 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ec1a91-8224-4f4e-a4ed-d0ee907b0354" path="/var/lib/kubelet/pods/53ec1a91-8224-4f4e-a4ed-d0ee907b0354/volumes" Mar 20 11:45:15 crc kubenswrapper[4846]: I0320 11:45:15.578389 4846 scope.go:117] "RemoveContainer" containerID="5203b292d231b364464cd68a9277c86c387901b379f9a0fde9581e81195e5c86" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.141753 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566786-zjpgg"] Mar 20 11:46:00 crc kubenswrapper[4846]: E0320 11:46:00.142875 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" containerName="collect-profiles" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.142894 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" containerName="collect-profiles" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.143081 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8768aa20-2e5f-4a04-a79d-81deb3f9dc3e" containerName="collect-profiles" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.143585 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.145746 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.150955 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.152790 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566786-zjpgg"] Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.153496 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.219571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xnzj\" (UniqueName: \"kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj\") pod \"auto-csr-approver-29566786-zjpgg\" (UID: \"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2\") " pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.321812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xnzj\" (UniqueName: \"kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj\") pod \"auto-csr-approver-29566786-zjpgg\" (UID: \"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2\") " pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.345604 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xnzj\" (UniqueName: \"kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj\") pod \"auto-csr-approver-29566786-zjpgg\" (UID: \"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2\") " pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:00 crc kubenswrapper[4846]: I0320 11:46:00.467723 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:01 crc kubenswrapper[4846]: I0320 11:46:01.017553 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566786-zjpgg"] Mar 20 11:46:01 crc kubenswrapper[4846]: I0320 11:46:01.024505 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:46:01 crc kubenswrapper[4846]: I0320 11:46:01.247702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" event={"ID":"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2","Type":"ContainerStarted","Data":"67fc847bbccc05448940bffa1369555a278703f0fe1dea93596c25e0edf0a789"} Mar 20 11:46:02 crc kubenswrapper[4846]: I0320 11:46:02.257607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" event={"ID":"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2","Type":"ContainerStarted","Data":"b420ff8e526514cbd95a25ef0482ec0e4c55f12daf2cb2161f4e455dc83d32f1"} Mar 20 11:46:02 crc kubenswrapper[4846]: I0320 11:46:02.276769 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" podStartSLOduration=1.324690431 podStartE2EDuration="2.276750189s" podCreationTimestamp="2026-03-20 11:46:00 +0000 UTC" firstStartedPulling="2026-03-20 11:46:01.024286804 +0000 UTC m=+2932.390695040" lastFinishedPulling="2026-03-20 11:46:01.976346562 +0000 UTC m=+2933.342754798" observedRunningTime="2026-03-20 11:46:02.274735729 +0000 UTC m=+2933.641143965" watchObservedRunningTime="2026-03-20 11:46:02.276750189 +0000 UTC m=+2933.643158425" Mar 20 11:46:03 crc kubenswrapper[4846]: I0320 11:46:03.268231 4846 generic.go:334] "Generic (PLEG): container finished" podID="db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" containerID="b420ff8e526514cbd95a25ef0482ec0e4c55f12daf2cb2161f4e455dc83d32f1" exitCode=0 Mar 20 11:46:03 crc kubenswrapper[4846]: I0320 11:46:03.268727 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" event={"ID":"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2","Type":"ContainerDied","Data":"b420ff8e526514cbd95a25ef0482ec0e4c55f12daf2cb2161f4e455dc83d32f1"} Mar 20 11:46:04 crc kubenswrapper[4846]: I0320 11:46:04.566966 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:04 crc kubenswrapper[4846]: I0320 11:46:04.689398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xnzj\" (UniqueName: \"kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj\") pod \"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2\" (UID: \"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2\") " Mar 20 11:46:04 crc kubenswrapper[4846]: I0320 11:46:04.697182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj" (OuterVolumeSpecName: "kube-api-access-9xnzj") pod "db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" (UID: "db43ed65-2a0a-4ed3-b37b-dea97fb65cd2"). InnerVolumeSpecName "kube-api-access-9xnzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:46:04 crc kubenswrapper[4846]: I0320 11:46:04.791590 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xnzj\" (UniqueName: \"kubernetes.io/projected/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2-kube-api-access-9xnzj\") on node \"crc\" DevicePath \"\"" Mar 20 11:46:05 crc kubenswrapper[4846]: I0320 11:46:05.285067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" event={"ID":"db43ed65-2a0a-4ed3-b37b-dea97fb65cd2","Type":"ContainerDied","Data":"67fc847bbccc05448940bffa1369555a278703f0fe1dea93596c25e0edf0a789"} Mar 20 11:46:05 crc kubenswrapper[4846]: I0320 11:46:05.285135 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67fc847bbccc05448940bffa1369555a278703f0fe1dea93596c25e0edf0a789" Mar 20 11:46:05 crc kubenswrapper[4846]: I0320 11:46:05.285207 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566786-zjpgg" Mar 20 11:46:05 crc kubenswrapper[4846]: I0320 11:46:05.354632 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566780-gxgf4"] Mar 20 11:46:05 crc kubenswrapper[4846]: I0320 11:46:05.360167 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566780-gxgf4"] Mar 20 11:46:07 crc kubenswrapper[4846]: I0320 11:46:07.336042 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63aac89a-b859-4734-8c82-0fd5a7bd373a" path="/var/lib/kubelet/pods/63aac89a-b859-4734-8c82-0fd5a7bd373a/volumes" Mar 20 11:46:15 crc kubenswrapper[4846]: I0320 11:46:15.637848 4846 scope.go:117] "RemoveContainer" containerID="a22a3f36a3684fe48674376a80e47d3ae371dd01c3b0392123b9dfcbc296d662" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.139248 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:31 crc kubenswrapper[4846]: E0320 11:46:31.140411 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" containerName="oc" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.140434 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" containerName="oc" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.140580 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" containerName="oc" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.141683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.150326 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.317143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.317221 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.317246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-679fz\" (UniqueName: \"kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.419487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.419535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-679fz\" (UniqueName: \"kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.419720 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.420205 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.420791 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.440726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-679fz\" (UniqueName: \"kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz\") pod \"redhat-operators-dqjfz\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.464143 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:31 crc kubenswrapper[4846]: I0320 11:46:31.706029 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:32 crc kubenswrapper[4846]: I0320 11:46:32.499041 4846 generic.go:334] "Generic (PLEG): container finished" podID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerID="039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd" exitCode=0 Mar 20 11:46:32 crc kubenswrapper[4846]: I0320 11:46:32.499091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerDied","Data":"039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd"} Mar 20 11:46:32 crc kubenswrapper[4846]: I0320 11:46:32.499116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerStarted","Data":"2a6a8407143016e09ae2ce92e9e47ddeae56d49d7a340770391009c08b14767d"} Mar 20 11:46:33 crc kubenswrapper[4846]: I0320 11:46:33.508783 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerStarted","Data":"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6"} Mar 20 11:46:34 crc kubenswrapper[4846]: I0320 11:46:34.517769 4846 generic.go:334] "Generic (PLEG): container finished" podID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerID="269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6" exitCode=0 Mar 20 11:46:34 crc kubenswrapper[4846]: I0320 11:46:34.517825 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerDied","Data":"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6"} Mar 20 11:46:35 crc kubenswrapper[4846]: I0320 11:46:35.527102 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerStarted","Data":"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274"} Mar 20 11:46:35 crc kubenswrapper[4846]: I0320 11:46:35.551509 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqjfz" podStartSLOduration=1.990067425 podStartE2EDuration="4.551489361s" podCreationTimestamp="2026-03-20 11:46:31 +0000 UTC" firstStartedPulling="2026-03-20 11:46:32.500628858 +0000 UTC m=+2963.867037094" lastFinishedPulling="2026-03-20 11:46:35.062050784 +0000 UTC m=+2966.428459030" observedRunningTime="2026-03-20 11:46:35.54696683 +0000 UTC m=+2966.913375076" watchObservedRunningTime="2026-03-20 11:46:35.551489361 +0000 UTC m=+2966.917897597" Mar 20 11:46:39 crc kubenswrapper[4846]: I0320 11:46:39.677480 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:46:39 crc kubenswrapper[4846]: I0320 11:46:39.677996 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:46:41 crc kubenswrapper[4846]: I0320 11:46:41.464471 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:41 crc kubenswrapper[4846]: I0320 11:46:41.465011 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:41 crc kubenswrapper[4846]: I0320 11:46:41.509359 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:41 crc kubenswrapper[4846]: I0320 11:46:41.616683 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:41 crc kubenswrapper[4846]: I0320 11:46:41.746228 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:43 crc kubenswrapper[4846]: I0320 11:46:43.593972 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqjfz" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="registry-server" containerID="cri-o://77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274" gracePeriod=2 Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.000518 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.108216 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-679fz\" (UniqueName: \"kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz\") pod \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.108356 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content\") pod \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.108510 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities\") pod \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\" (UID: \"53479767-ed9d-4f8a-b7e7-a01176fb8c41\") " Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.109754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities" (OuterVolumeSpecName: "utilities") pod "53479767-ed9d-4f8a-b7e7-a01176fb8c41" (UID: "53479767-ed9d-4f8a-b7e7-a01176fb8c41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.117150 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz" (OuterVolumeSpecName: "kube-api-access-679fz") pod "53479767-ed9d-4f8a-b7e7-a01176fb8c41" (UID: "53479767-ed9d-4f8a-b7e7-a01176fb8c41"). InnerVolumeSpecName "kube-api-access-679fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.210024 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.210076 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-679fz\" (UniqueName: \"kubernetes.io/projected/53479767-ed9d-4f8a-b7e7-a01176fb8c41-kube-api-access-679fz\") on node \"crc\" DevicePath \"\"" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.608700 4846 generic.go:334] "Generic (PLEG): container finished" podID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerID="77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274" exitCode=0 Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.608778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerDied","Data":"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274"} Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.610298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqjfz" event={"ID":"53479767-ed9d-4f8a-b7e7-a01176fb8c41","Type":"ContainerDied","Data":"2a6a8407143016e09ae2ce92e9e47ddeae56d49d7a340770391009c08b14767d"} Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.610330 4846 scope.go:117] "RemoveContainer" containerID="77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.608828 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqjfz" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.631938 4846 scope.go:117] "RemoveContainer" containerID="269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.657425 4846 scope.go:117] "RemoveContainer" containerID="039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.703919 4846 scope.go:117] "RemoveContainer" containerID="77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274" Mar 20 11:46:44 crc kubenswrapper[4846]: E0320 11:46:44.704384 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274\": container with ID starting with 77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274 not found: ID does not exist" containerID="77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.704419 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274"} err="failed to get container status \"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274\": rpc error: code = NotFound desc = could not find container \"77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274\": container with ID starting with 77f46bcea9aefe9737e957a380ec7323e474e1214bdf97b64ace59785bb46274 not found: ID does not exist" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.704514 4846 scope.go:117] "RemoveContainer" containerID="269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6" Mar 20 11:46:44 crc kubenswrapper[4846]: E0320 11:46:44.705088 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6\": container with ID starting with 269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6 not found: ID does not exist" containerID="269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.705121 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6"} err="failed to get container status \"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6\": rpc error: code = NotFound desc = could not find container \"269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6\": container with ID starting with 269d324e744dca952e40ce151a4f9b65a05dd1ebf12e2e232ea16cf9857255b6 not found: ID does not exist" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.705141 4846 scope.go:117] "RemoveContainer" containerID="039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd" Mar 20 11:46:44 crc kubenswrapper[4846]: E0320 11:46:44.705501 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd\": container with ID starting with 039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd not found: ID does not exist" containerID="039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd" Mar 20 11:46:44 crc kubenswrapper[4846]: I0320 11:46:44.705531 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd"} err="failed to get container status \"039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd\": rpc error: code = NotFound desc = could not find container \"039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd\": container with ID starting with 039ad899d7ea696f249b77a76541a309167a25765173dc3eeb9395118adf54fd not found: ID does not exist" Mar 20 11:46:46 crc kubenswrapper[4846]: I0320 11:46:46.378645 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53479767-ed9d-4f8a-b7e7-a01176fb8c41" (UID: "53479767-ed9d-4f8a-b7e7-a01176fb8c41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:46:46 crc kubenswrapper[4846]: I0320 11:46:46.442992 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53479767-ed9d-4f8a-b7e7-a01176fb8c41-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:46:46 crc kubenswrapper[4846]: I0320 11:46:46.449371 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:46 crc kubenswrapper[4846]: I0320 11:46:46.455161 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqjfz"] Mar 20 11:46:47 crc kubenswrapper[4846]: I0320 11:46:47.332365 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" path="/var/lib/kubelet/pods/53479767-ed9d-4f8a-b7e7-a01176fb8c41/volumes" Mar 20 11:47:09 crc kubenswrapper[4846]: I0320 11:47:09.677750 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:47:09 crc kubenswrapper[4846]: I0320 11:47:09.678645 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.048853 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lhdz8/must-gather-gkdck"] Mar 20 11:47:10 crc kubenswrapper[4846]: E0320 11:47:10.049729 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="extract-content" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.049756 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="extract-content" Mar 20 11:47:10 crc kubenswrapper[4846]: E0320 11:47:10.049779 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="extract-utilities" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.049790 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="extract-utilities" Mar 20 11:47:10 crc kubenswrapper[4846]: E0320 11:47:10.049806 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="registry-server" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.049814 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="registry-server" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.049984 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="53479767-ed9d-4f8a-b7e7-a01176fb8c41" containerName="registry-server" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.050818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.058034 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lhdz8"/"openshift-service-ca.crt" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.058138 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lhdz8"/"kube-root-ca.crt" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.058213 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lhdz8"/"default-dockercfg-rrk28" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.079798 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lhdz8/must-gather-gkdck"] Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.138259 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.138571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhk9v\" (UniqueName: \"kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.240209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.240319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhk9v\" (UniqueName: \"kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.241038 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.264421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhk9v\" (UniqueName: \"kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v\") pod \"must-gather-gkdck\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.379926 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:47:10 crc kubenswrapper[4846]: I0320 11:47:10.845080 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lhdz8/must-gather-gkdck"] Mar 20 11:47:11 crc kubenswrapper[4846]: I0320 11:47:11.818509 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lhdz8/must-gather-gkdck" event={"ID":"fb3fbb06-e0a8-46a1-8f80-565f59acb451","Type":"ContainerStarted","Data":"0b89702860eec9ee1849f9159833bbc60111d784a614c3a2d68c03cebc3bff9d"} Mar 20 11:47:19 crc kubenswrapper[4846]: I0320 11:47:19.898335 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lhdz8/must-gather-gkdck" event={"ID":"fb3fbb06-e0a8-46a1-8f80-565f59acb451","Type":"ContainerStarted","Data":"fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad"} Mar 20 11:47:19 crc kubenswrapper[4846]: I0320 11:47:19.899141 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lhdz8/must-gather-gkdck" event={"ID":"fb3fbb06-e0a8-46a1-8f80-565f59acb451","Type":"ContainerStarted","Data":"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31"} Mar 20 11:47:39 crc kubenswrapper[4846]: I0320 11:47:39.678094 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:47:39 crc kubenswrapper[4846]: I0320 11:47:39.678880 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:47:39 crc kubenswrapper[4846]: I0320 11:47:39.678956 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:47:39 crc kubenswrapper[4846]: I0320 11:47:39.679642 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:47:39 crc kubenswrapper[4846]: I0320 11:47:39.679704 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" gracePeriod=600 Mar 20 11:47:39 crc kubenswrapper[4846]: E0320 11:47:39.825479 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:47:40 crc kubenswrapper[4846]: I0320 11:47:40.542832 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" exitCode=0 Mar 20 11:47:40 crc kubenswrapper[4846]: I0320 11:47:40.542885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60"} Mar 20 11:47:40 crc kubenswrapper[4846]: I0320 11:47:40.542980 4846 scope.go:117] "RemoveContainer" containerID="b587463a47da40507cdce2f8e4b237c353cf8c5b47d6e8b6f6998792e41e3db4" Mar 20 11:47:40 crc kubenswrapper[4846]: I0320 11:47:40.543644 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:47:40 crc kubenswrapper[4846]: E0320 11:47:40.543998 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:47:40 crc kubenswrapper[4846]: I0320 11:47:40.569186 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lhdz8/must-gather-gkdck" podStartSLOduration=22.323789076 podStartE2EDuration="30.569156918s" podCreationTimestamp="2026-03-20 11:47:10 +0000 UTC" firstStartedPulling="2026-03-20 11:47:10.861055326 +0000 UTC m=+3002.227463562" lastFinishedPulling="2026-03-20 11:47:19.106423168 +0000 UTC m=+3010.472831404" observedRunningTime="2026-03-20 11:47:19.929912102 +0000 UTC m=+3011.296320338" watchObservedRunningTime="2026-03-20 11:47:40.569156918 +0000 UTC m=+3031.935565154" Mar 20 11:47:52 crc kubenswrapper[4846]: I0320 11:47:52.322684 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:47:52 crc kubenswrapper[4846]: E0320 11:47:52.323844 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.143692 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566788-8m5k9"] Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.145669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.147714 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.148132 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.148643 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.155973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566788-8m5k9"] Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.228682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn9sn\" (UniqueName: \"kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn\") pod \"auto-csr-approver-29566788-8m5k9\" (UID: \"103af8a1-8d20-47da-91d8-168573c714b3\") " pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.330493 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn9sn\" (UniqueName: \"kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn\") pod \"auto-csr-approver-29566788-8m5k9\" (UID: \"103af8a1-8d20-47da-91d8-168573c714b3\") " pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.351876 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn9sn\" (UniqueName: \"kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn\") pod \"auto-csr-approver-29566788-8m5k9\" (UID: \"103af8a1-8d20-47da-91d8-168573c714b3\") " pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.467530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:00 crc kubenswrapper[4846]: I0320 11:48:00.954103 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566788-8m5k9"] Mar 20 11:48:01 crc kubenswrapper[4846]: I0320 11:48:01.702738 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" event={"ID":"103af8a1-8d20-47da-91d8-168573c714b3","Type":"ContainerStarted","Data":"569252cac740c2586fc94aa95c322b6f99af7c8aeb6d81df7514869db863b4f6"} Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.153222 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.155657 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.164016 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.166570 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.166878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb9z5\" (UniqueName: \"kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.167061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.268340 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb9z5\" (UniqueName: \"kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.268403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.268448 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.269032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.269455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.294283 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb9z5\" (UniqueName: \"kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5\") pod \"community-operators-5xfbm\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.480004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.730704 4846 generic.go:334] "Generic (PLEG): container finished" podID="103af8a1-8d20-47da-91d8-168573c714b3" containerID="f5843a287dcfa07c01224499ee4c3faa841556b2db84f82bf3fa86ddb5d18cf5" exitCode=0 Mar 20 11:48:02 crc kubenswrapper[4846]: I0320 11:48:02.730821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" event={"ID":"103af8a1-8d20-47da-91d8-168573c714b3","Type":"ContainerDied","Data":"f5843a287dcfa07c01224499ee4c3faa841556b2db84f82bf3fa86ddb5d18cf5"} Mar 20 11:48:03 crc kubenswrapper[4846]: I0320 11:48:03.012876 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:03 crc kubenswrapper[4846]: W0320 11:48:03.017116 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09e81b17_838f_41b6_87d9_52da8d494d71.slice/crio-9946d82f06c7b43387abd47e43f5bc43377f6975ae7af34991fc90f3476090d3 WatchSource:0}: Error finding container 9946d82f06c7b43387abd47e43f5bc43377f6975ae7af34991fc90f3476090d3: Status 404 returned error can't find the container with id 9946d82f06c7b43387abd47e43f5bc43377f6975ae7af34991fc90f3476090d3 Mar 20 11:48:03 crc kubenswrapper[4846]: I0320 11:48:03.323021 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:48:03 crc kubenswrapper[4846]: E0320 11:48:03.323718 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:03 crc kubenswrapper[4846]: I0320 11:48:03.753163 4846 generic.go:334] "Generic (PLEG): container finished" podID="09e81b17-838f-41b6-87d9-52da8d494d71" containerID="52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da" exitCode=0 Mar 20 11:48:03 crc kubenswrapper[4846]: I0320 11:48:03.753285 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerDied","Data":"52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da"} Mar 20 11:48:03 crc kubenswrapper[4846]: I0320 11:48:03.754030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerStarted","Data":"9946d82f06c7b43387abd47e43f5bc43377f6975ae7af34991fc90f3476090d3"} Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.041525 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.197409 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn9sn\" (UniqueName: \"kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn\") pod \"103af8a1-8d20-47da-91d8-168573c714b3\" (UID: \"103af8a1-8d20-47da-91d8-168573c714b3\") " Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.204131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn" (OuterVolumeSpecName: "kube-api-access-mn9sn") pod "103af8a1-8d20-47da-91d8-168573c714b3" (UID: "103af8a1-8d20-47da-91d8-168573c714b3"). InnerVolumeSpecName "kube-api-access-mn9sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.299259 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn9sn\" (UniqueName: \"kubernetes.io/projected/103af8a1-8d20-47da-91d8-168573c714b3-kube-api-access-mn9sn\") on node \"crc\" DevicePath \"\"" Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.762887 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.766968 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566788-8m5k9" event={"ID":"103af8a1-8d20-47da-91d8-168573c714b3","Type":"ContainerDied","Data":"569252cac740c2586fc94aa95c322b6f99af7c8aeb6d81df7514869db863b4f6"} Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.767017 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569252cac740c2586fc94aa95c322b6f99af7c8aeb6d81df7514869db863b4f6" Mar 20 11:48:04 crc kubenswrapper[4846]: I0320 11:48:04.770542 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerStarted","Data":"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0"} Mar 20 11:48:05 crc kubenswrapper[4846]: I0320 11:48:05.118709 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566782-glcd2"] Mar 20 11:48:05 crc kubenswrapper[4846]: I0320 11:48:05.128838 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566782-glcd2"] Mar 20 11:48:05 crc kubenswrapper[4846]: I0320 11:48:05.331308 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29560ab-925d-4746-90ea-fb3f295e90c0" path="/var/lib/kubelet/pods/d29560ab-925d-4746-90ea-fb3f295e90c0/volumes" Mar 20 11:48:05 crc kubenswrapper[4846]: I0320 11:48:05.781384 4846 generic.go:334] "Generic (PLEG): container finished" podID="09e81b17-838f-41b6-87d9-52da8d494d71" containerID="3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0" exitCode=0 Mar 20 11:48:05 crc kubenswrapper[4846]: I0320 11:48:05.781439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerDied","Data":"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0"} Mar 20 11:48:07 crc kubenswrapper[4846]: I0320 11:48:07.725404 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-q2qnv_e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4/init/0.log" Mar 20 11:48:07 crc kubenswrapper[4846]: I0320 11:48:07.798837 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerStarted","Data":"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1"} Mar 20 11:48:07 crc kubenswrapper[4846]: I0320 11:48:07.817982 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5xfbm" podStartSLOduration=2.925733772 podStartE2EDuration="5.817956796s" podCreationTimestamp="2026-03-20 11:48:02 +0000 UTC" firstStartedPulling="2026-03-20 11:48:03.754732725 +0000 UTC m=+3055.121140951" lastFinishedPulling="2026-03-20 11:48:06.646955739 +0000 UTC m=+3058.013363975" observedRunningTime="2026-03-20 11:48:07.815638069 +0000 UTC m=+3059.182046315" watchObservedRunningTime="2026-03-20 11:48:07.817956796 +0000 UTC m=+3059.184365032" Mar 20 11:48:07 crc kubenswrapper[4846]: I0320 11:48:07.903729 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-q2qnv_e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4/dnsmasq-dns/0.log" Mar 20 11:48:07 crc kubenswrapper[4846]: I0320 11:48:07.948987 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-q2qnv_e8dacc6e-4efd-4782-b0dd-f942e7b9bfc4/init/0.log" Mar 20 11:48:12 crc kubenswrapper[4846]: I0320 11:48:12.480483 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:12 crc kubenswrapper[4846]: I0320 11:48:12.481329 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:12 crc kubenswrapper[4846]: I0320 11:48:12.546592 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:12 crc kubenswrapper[4846]: I0320 11:48:12.898297 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:12 crc kubenswrapper[4846]: I0320 11:48:12.956692 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:14 crc kubenswrapper[4846]: I0320 11:48:14.877859 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5xfbm" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="registry-server" containerID="cri-o://1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1" gracePeriod=2 Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.287766 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.399752 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities\") pod \"09e81b17-838f-41b6-87d9-52da8d494d71\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.399841 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content\") pod \"09e81b17-838f-41b6-87d9-52da8d494d71\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.399972 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb9z5\" (UniqueName: \"kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5\") pod \"09e81b17-838f-41b6-87d9-52da8d494d71\" (UID: \"09e81b17-838f-41b6-87d9-52da8d494d71\") " Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.405767 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities" (OuterVolumeSpecName: "utilities") pod "09e81b17-838f-41b6-87d9-52da8d494d71" (UID: "09e81b17-838f-41b6-87d9-52da8d494d71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.412923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5" (OuterVolumeSpecName: "kube-api-access-hb9z5") pod "09e81b17-838f-41b6-87d9-52da8d494d71" (UID: "09e81b17-838f-41b6-87d9-52da8d494d71"). InnerVolumeSpecName "kube-api-access-hb9z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.501957 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb9z5\" (UniqueName: \"kubernetes.io/projected/09e81b17-838f-41b6-87d9-52da8d494d71-kube-api-access-hb9z5\") on node \"crc\" DevicePath \"\"" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.501991 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.887295 4846 generic.go:334] "Generic (PLEG): container finished" podID="09e81b17-838f-41b6-87d9-52da8d494d71" containerID="1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1" exitCode=0 Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.887351 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerDied","Data":"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1"} Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.887370 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xfbm" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.887389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xfbm" event={"ID":"09e81b17-838f-41b6-87d9-52da8d494d71","Type":"ContainerDied","Data":"9946d82f06c7b43387abd47e43f5bc43377f6975ae7af34991fc90f3476090d3"} Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.887411 4846 scope.go:117] "RemoveContainer" containerID="1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.907295 4846 scope.go:117] "RemoveContainer" containerID="3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.926359 4846 scope.go:117] "RemoveContainer" containerID="52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.950197 4846 scope.go:117] "RemoveContainer" containerID="1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1" Mar 20 11:48:15 crc kubenswrapper[4846]: E0320 11:48:15.950735 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1\": container with ID starting with 1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1 not found: ID does not exist" containerID="1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.950765 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1"} err="failed to get container status \"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1\": rpc error: code = NotFound desc = could not find container \"1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1\": container with ID starting with 1960afd733011e17c4eee7e5e9da453d26e32e02e493da3284fa328eab0957b1 not found: ID does not exist" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.950786 4846 scope.go:117] "RemoveContainer" containerID="3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0" Mar 20 11:48:15 crc kubenswrapper[4846]: E0320 11:48:15.951159 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0\": container with ID starting with 3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0 not found: ID does not exist" containerID="3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.951180 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0"} err="failed to get container status \"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0\": rpc error: code = NotFound desc = could not find container \"3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0\": container with ID starting with 3ea3e6c3574c0dc8e2f88f3bb4aa576b5e597479489b9361ca720b09d4f672a0 not found: ID does not exist" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.951198 4846 scope.go:117] "RemoveContainer" containerID="52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da" Mar 20 11:48:15 crc kubenswrapper[4846]: E0320 11:48:15.951580 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da\": container with ID starting with 52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da not found: ID does not exist" containerID="52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da" Mar 20 11:48:15 crc kubenswrapper[4846]: I0320 11:48:15.951604 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da"} err="failed to get container status \"52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da\": rpc error: code = NotFound desc = could not find container \"52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da\": container with ID starting with 52b67471dd7f79e4b876fbdf09b11d8fd1d981e5955226b727f7ce092eba37da not found: ID does not exist" Mar 20 11:48:16 crc kubenswrapper[4846]: I0320 11:48:16.151528 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09e81b17-838f-41b6-87d9-52da8d494d71" (UID: "09e81b17-838f-41b6-87d9-52da8d494d71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:48:16 crc kubenswrapper[4846]: I0320 11:48:16.211921 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e81b17-838f-41b6-87d9-52da8d494d71-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:48:16 crc kubenswrapper[4846]: I0320 11:48:16.220584 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:16 crc kubenswrapper[4846]: I0320 11:48:16.231008 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5xfbm"] Mar 20 11:48:16 crc kubenswrapper[4846]: I0320 11:48:16.322797 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:48:16 crc kubenswrapper[4846]: E0320 11:48:16.323283 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:17 crc kubenswrapper[4846]: I0320 11:48:17.333875 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" path="/var/lib/kubelet/pods/09e81b17-838f-41b6-87d9-52da8d494d71/volumes" Mar 20 11:48:18 crc kubenswrapper[4846]: I0320 11:48:18.664538 4846 scope.go:117] "RemoveContainer" containerID="92c11d91c7ed5cfa83ec55250b78a8d26e87b63a161a597d4942ce5b78d99c42" Mar 20 11:48:22 crc kubenswrapper[4846]: I0320 11:48:22.623822 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/util/0.log" Mar 20 11:48:22 crc kubenswrapper[4846]: I0320 11:48:22.801577 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/pull/0.log" Mar 20 11:48:22 crc kubenswrapper[4846]: I0320 11:48:22.865621 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/util/0.log" Mar 20 11:48:22 crc kubenswrapper[4846]: I0320 11:48:22.885796 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/pull/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.076450 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/pull/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.099507 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/util/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.140572 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_54f8d1291b8fed3b88b1c26a19da77e064028ff70927209c4420d1682a4tvjr_c4028a92-49ba-4337-b2a7-e15686ebf8eb/extract/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.273087 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-zncs8_529816a9-1274-4a4f-88b2-6e0b426075a0/manager/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.538339 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-ht246_9bf60e8a-5221-44ba-8471-5b9e2692d2a7/manager/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.632001 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-4m4lg_380b5a6a-c151-4763-a86b-c8e53b6f1aa2/manager/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.818073 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-thhs6_24727b45-3896-4b9f-a91e-62434aa32ac7/manager/0.log" Mar 20 11:48:23 crc kubenswrapper[4846]: I0320 11:48:23.966840 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-ggrpk_aff6e349-68ea-4f3c-8514-b2a7c7ffed23/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.103143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-qg5jf_6be091f8-9109-4fce-a856-feacf04c76d4/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.118456 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-594f6f699b-vmg46_1132b3dc-feeb-41ce-a579-d5b089b86597/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.255293 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-g8hzn_c88c55bf-7869-40e5-90f7-0bfdcb676b9e/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.338292 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-qwptc_2bba6588-37b3-453e-972e-393208ddd560/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.496403 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-hgqt9_86d13cbe-cc30-41b6-8829-1f5cbb99c48c/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.574707 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-r7pgd_0b70e378-a53e-4519-b415-64810a544874/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.715842 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-9wzfn_fa16f94f-be9c-4e2c-8dc3-c420db3ed639/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.780727 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-f45j6_bd21bfeb-d2b2-4757-aa36-094c039951f0/manager/0.log" Mar 20 11:48:24 crc kubenswrapper[4846]: I0320 11:48:24.928020 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-rrx9r_06eb056c-f3f0-4242-bbee-d201dfd37925/manager/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.187170 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-46rnd_7d330820-2aac-4076-a205-8dedd331cae1/manager/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.455719 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7575fb585b-v27cw_de215fcd-2be6-40ca-8ae1-d628b6b4280b/manager/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.468887 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-7fc6f9b5fc-djp94_4b359265-9a42-4e56-a6ea-108fafb33ae3/operator/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.648527 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hkt64_bc323795-0c0f-4463-9b41-f62b2278713a/registry-server/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.710764 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-gqbrp_64ce367c-20db-4598-8006-f51d8aa22bdd/manager/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.894131 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-dr5ms_132090d4-b1d3-485e-9525-26fe1eecc448/manager/0.log" Mar 20 11:48:25 crc kubenswrapper[4846]: I0320 11:48:25.950017 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-4fww7_9c7c760d-4ef7-41a0-bca6-c9b26d99fd52/manager/0.log" Mar 20 11:48:26 crc kubenswrapper[4846]: I0320 11:48:26.109565 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-ppjfc_14ce4a99-f370-4aaa-bb8f-bd3f1d8746dd/manager/0.log" Mar 20 11:48:26 crc kubenswrapper[4846]: I0320 11:48:26.228124 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-b8ng9_e7f86e9b-89bb-4558-aa0d-4bec37131052/manager/0.log" Mar 20 11:48:26 crc kubenswrapper[4846]: I0320 11:48:26.362384 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-dvj5j_b155d1b3-b553-414b-9017-329ee45a9658/manager/0.log" Mar 20 11:48:27 crc kubenswrapper[4846]: I0320 11:48:27.327073 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:48:27 crc kubenswrapper[4846]: E0320 11:48:27.327284 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:38 crc kubenswrapper[4846]: I0320 11:48:38.323439 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:48:38 crc kubenswrapper[4846]: E0320 11:48:38.324487 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:46 crc kubenswrapper[4846]: I0320 11:48:46.063697 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sv5bj_779260d5-a4b3-41d5-9a65-86dbb0c66845/control-plane-machine-set-operator/0.log" Mar 20 11:48:46 crc kubenswrapper[4846]: I0320 11:48:46.286305 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t4dz2_36e90b7b-0251-46ef-8ae6-b5522e3138d2/kube-rbac-proxy/0.log" Mar 20 11:48:46 crc kubenswrapper[4846]: I0320 11:48:46.304412 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t4dz2_36e90b7b-0251-46ef-8ae6-b5522e3138d2/machine-api-operator/0.log" Mar 20 11:48:52 crc kubenswrapper[4846]: I0320 11:48:52.322828 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:48:52 crc kubenswrapper[4846]: E0320 11:48:52.323836 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:48:58 crc kubenswrapper[4846]: I0320 11:48:58.987427 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-dwsx8_64fa0d45-f6a9-4b1f-8381-8d0cf63780e9/cert-manager-controller/0.log" Mar 20 11:48:59 crc kubenswrapper[4846]: I0320 11:48:59.179090 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xwvmq_9480649a-6163-4ab6-90be-d825b1d39724/cert-manager-cainjector/0.log" Mar 20 11:48:59 crc kubenswrapper[4846]: I0320 11:48:59.263606 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-dcs94_0794f7a2-b0b8-419a-b7db-eb75a2dcbf73/cert-manager-webhook/0.log" Mar 20 11:49:03 crc kubenswrapper[4846]: I0320 11:49:03.322858 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:49:03 crc kubenswrapper[4846]: E0320 11:49:03.325385 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.169997 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-mpkwp_7c757d6d-23d9-4555-8f78-c9bd05130eae/nmstate-console-plugin/0.log" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.337526 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-cl5nt_0c693da7-8768-44e0-a24a-d1f770cf0e4a/nmstate-handler/0.log" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.438689 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-pzj9s_8bdeb909-49d8-4de7-98dc-ac1f1466dc98/kube-rbac-proxy/0.log" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.444565 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-pzj9s_8bdeb909-49d8-4de7-98dc-ac1f1466dc98/nmstate-metrics/0.log" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.607142 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-fcc6h_5db89f9f-d51b-430f-88de-34f48c693c58/nmstate-operator/0.log" Mar 20 11:49:11 crc kubenswrapper[4846]: I0320 11:49:11.676880 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-pbzwm_13c7db72-fd6b-4cdc-998a-163bec94b0f3/nmstate-webhook/0.log" Mar 20 11:49:16 crc kubenswrapper[4846]: I0320 11:49:16.322407 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:49:16 crc kubenswrapper[4846]: E0320 11:49:16.323385 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:49:28 crc kubenswrapper[4846]: I0320 11:49:28.323115 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:49:28 crc kubenswrapper[4846]: E0320 11:49:28.323912 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:49:37 crc kubenswrapper[4846]: I0320 11:49:37.965845 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-hvrww_bc262413-3fc0-4204-a662-b8e3c1a9eb15/kube-rbac-proxy/0.log" Mar 20 11:49:37 crc kubenswrapper[4846]: I0320 11:49:37.977083 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-hvrww_bc262413-3fc0-4204-a662-b8e3c1a9eb15/controller/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.169097 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-frr-files/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.613648 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-metrics/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.615534 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-frr-files/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.634151 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-reloader/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.670242 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-reloader/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.867220 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-frr-files/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.881979 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-metrics/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.887494 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-reloader/0.log" Mar 20 11:49:38 crc kubenswrapper[4846]: I0320 11:49:38.909308 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-metrics/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.113793 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-reloader/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.144251 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/controller/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.189739 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-frr-files/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.191288 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/cp-metrics/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.337470 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:49:39 crc kubenswrapper[4846]: E0320 11:49:39.337721 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.361398 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/frr-metrics/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.378706 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/kube-rbac-proxy-frr/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.400931 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/kube-rbac-proxy/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.572406 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/reloader/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.656417 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-d554q_e6c14c61-393f-4234-9d0f-bcc3cf725c6e/frr/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.703135 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-96pp8_98b51565-aa74-437c-9d42-a93f2b6b52de/frr-k8s-webhook-server/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.874929 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66844494d6-8t8lj_78d82817-160a-4eb0-96ba-da47a62bd61c/manager/0.log" Mar 20 11:49:39 crc kubenswrapper[4846]: I0320 11:49:39.907393 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-cd87fcd9-tfmhq_5a3417e0-ef3f-461d-b798-caab9527c70b/webhook-server/0.log" Mar 20 11:49:40 crc kubenswrapper[4846]: I0320 11:49:40.076007 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jjwfh_08211b7a-f5ad-48aa-85ac-be288b9e3e40/kube-rbac-proxy/0.log" Mar 20 11:49:40 crc kubenswrapper[4846]: I0320 11:49:40.242286 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jjwfh_08211b7a-f5ad-48aa-85ac-be288b9e3e40/speaker/0.log" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.445433 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:41 crc kubenswrapper[4846]: E0320 11:49:41.445837 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="extract-content" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.445857 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="extract-content" Mar 20 11:49:41 crc kubenswrapper[4846]: E0320 11:49:41.445877 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="registry-server" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.445884 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="registry-server" Mar 20 11:49:41 crc kubenswrapper[4846]: E0320 11:49:41.445927 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="extract-utilities" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.445939 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="extract-utilities" Mar 20 11:49:41 crc kubenswrapper[4846]: E0320 11:49:41.445964 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103af8a1-8d20-47da-91d8-168573c714b3" containerName="oc" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.445972 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="103af8a1-8d20-47da-91d8-168573c714b3" containerName="oc" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.446139 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e81b17-838f-41b6-87d9-52da8d494d71" containerName="registry-server" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.446165 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="103af8a1-8d20-47da-91d8-168573c714b3" containerName="oc" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.447682 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.459757 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.598131 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.598257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.598292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kzvl\" (UniqueName: \"kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.704606 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.705118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kzvl\" (UniqueName: \"kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.705260 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.705406 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.705729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.732237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kzvl\" (UniqueName: \"kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl\") pod \"certified-operators-4jzhn\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:41 crc kubenswrapper[4846]: I0320 11:49:41.777701 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:42 crc kubenswrapper[4846]: I0320 11:49:42.323831 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:42 crc kubenswrapper[4846]: I0320 11:49:42.586273 4846 generic.go:334] "Generic (PLEG): container finished" podID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerID="d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899" exitCode=0 Mar 20 11:49:42 crc kubenswrapper[4846]: I0320 11:49:42.586325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerDied","Data":"d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899"} Mar 20 11:49:42 crc kubenswrapper[4846]: I0320 11:49:42.586373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerStarted","Data":"ad2a792c1d0bebd2dabd3f83a3301d30214552ae6350d244ca015064b434abe1"} Mar 20 11:49:43 crc kubenswrapper[4846]: I0320 11:49:43.597979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerStarted","Data":"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab"} Mar 20 11:49:44 crc kubenswrapper[4846]: I0320 11:49:44.606601 4846 generic.go:334] "Generic (PLEG): container finished" podID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerID="febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab" exitCode=0 Mar 20 11:49:44 crc kubenswrapper[4846]: I0320 11:49:44.606655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerDied","Data":"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab"} Mar 20 11:49:44 crc kubenswrapper[4846]: I0320 11:49:44.606681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerStarted","Data":"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888"} Mar 20 11:49:44 crc kubenswrapper[4846]: I0320 11:49:44.631529 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jzhn" podStartSLOduration=2.16299963 podStartE2EDuration="3.631499663s" podCreationTimestamp="2026-03-20 11:49:41 +0000 UTC" firstStartedPulling="2026-03-20 11:49:42.587994967 +0000 UTC m=+3153.954403203" lastFinishedPulling="2026-03-20 11:49:44.056495 +0000 UTC m=+3155.422903236" observedRunningTime="2026-03-20 11:49:44.629547885 +0000 UTC m=+3155.995956131" watchObservedRunningTime="2026-03-20 11:49:44.631499663 +0000 UTC m=+3155.997907899" Mar 20 11:49:51 crc kubenswrapper[4846]: I0320 11:49:51.779343 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:51 crc kubenswrapper[4846]: I0320 11:49:51.780200 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:51 crc kubenswrapper[4846]: I0320 11:49:51.835383 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:52 crc kubenswrapper[4846]: I0320 11:49:52.725199 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:53 crc kubenswrapper[4846]: I0320 11:49:53.323347 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:49:53 crc kubenswrapper[4846]: E0320 11:49:53.323819 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.367697 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/util/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.438522 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.559824 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/util/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.628675 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/pull/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.638969 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/pull/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.681685 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4jzhn" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="registry-server" containerID="cri-o://75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888" gracePeriod=2 Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.829334 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/pull/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.890133 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/extract/0.log" Mar 20 11:49:54 crc kubenswrapper[4846]: I0320 11:49:54.903245 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874qgwg7_eb990abc-5080-4db8-919a-e59085f1a8b0/util/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.095047 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/util/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.110109 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.233759 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content\") pod \"12fa44d8-e695-4275-b09f-7bae13d6476d\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.233858 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities\") pod \"12fa44d8-e695-4275-b09f-7bae13d6476d\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.233936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kzvl\" (UniqueName: \"kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl\") pod \"12fa44d8-e695-4275-b09f-7bae13d6476d\" (UID: \"12fa44d8-e695-4275-b09f-7bae13d6476d\") " Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.234975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities" (OuterVolumeSpecName: "utilities") pod "12fa44d8-e695-4275-b09f-7bae13d6476d" (UID: "12fa44d8-e695-4275-b09f-7bae13d6476d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.243106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl" (OuterVolumeSpecName: "kube-api-access-9kzvl") pod "12fa44d8-e695-4275-b09f-7bae13d6476d" (UID: "12fa44d8-e695-4275-b09f-7bae13d6476d"). InnerVolumeSpecName "kube-api-access-9kzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.338808 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kzvl\" (UniqueName: \"kubernetes.io/projected/12fa44d8-e695-4275-b09f-7bae13d6476d-kube-api-access-9kzvl\") on node \"crc\" DevicePath \"\"" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.338854 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.355629 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12fa44d8-e695-4275-b09f-7bae13d6476d" (UID: "12fa44d8-e695-4275-b09f-7bae13d6476d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.440651 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12fa44d8-e695-4275-b09f-7bae13d6476d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.526398 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/util/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.537057 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/pull/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.558850 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/pull/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.691836 4846 generic.go:334] "Generic (PLEG): container finished" podID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerID="75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888" exitCode=0 Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.691919 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jzhn" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.691907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerDied","Data":"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888"} Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.692492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jzhn" event={"ID":"12fa44d8-e695-4275-b09f-7bae13d6476d","Type":"ContainerDied","Data":"ad2a792c1d0bebd2dabd3f83a3301d30214552ae6350d244ca015064b434abe1"} Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.692520 4846 scope.go:117] "RemoveContainer" containerID="75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.729398 4846 scope.go:117] "RemoveContainer" containerID="febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.740981 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.748720 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/util/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.749434 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4jzhn"] Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.751791 4846 scope.go:117] "RemoveContainer" containerID="d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.768189 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/extract/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.772580 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1nz82d_a07c8578-7175-4eb7-9c40-416863db9cb2/pull/0.log" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.783632 4846 scope.go:117] "RemoveContainer" containerID="75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888" Mar 20 11:49:55 crc kubenswrapper[4846]: E0320 11:49:55.784111 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888\": container with ID starting with 75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888 not found: ID does not exist" containerID="75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.784148 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888"} err="failed to get container status \"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888\": rpc error: code = NotFound desc = could not find container \"75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888\": container with ID starting with 75701bd2b9925d2ac684b2d3c779deec9628dac2a9577b340a6d6849e440f888 not found: ID does not exist" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.784173 4846 scope.go:117] "RemoveContainer" containerID="febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab" Mar 20 11:49:55 crc kubenswrapper[4846]: E0320 11:49:55.784504 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab\": container with ID starting with febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab not found: ID does not exist" containerID="febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.784574 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab"} err="failed to get container status \"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab\": rpc error: code = NotFound desc = could not find container \"febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab\": container with ID starting with febd9f731bcbbb92ea092bfc88ac9dbb83638234f02a15e6b4882617371d52ab not found: ID does not exist" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.784610 4846 scope.go:117] "RemoveContainer" containerID="d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899" Mar 20 11:49:55 crc kubenswrapper[4846]: E0320 11:49:55.785138 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899\": container with ID starting with d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899 not found: ID does not exist" containerID="d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899" Mar 20 11:49:55 crc kubenswrapper[4846]: I0320 11:49:55.785177 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899"} err="failed to get container status \"d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899\": rpc error: code = NotFound desc = could not find container \"d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899\": container with ID starting with d7f456e6a457eddea6237bd79f6ce31045bb622900ba1e1d6a6e29b281880899 not found: ID does not exist" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.373287 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-utilities/0.log" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.543771 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-content/0.log" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.550438 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-utilities/0.log" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.572459 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-content/0.log" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.756563 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-utilities/0.log" Mar 20 11:49:56 crc kubenswrapper[4846]: I0320 11:49:56.800478 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/extract-content/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.029247 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-utilities/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.246873 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-content/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.286799 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7mkqk_4df61cc9-7ff8-47a8-9800-5f27801924f3/registry-server/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.327206 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-utilities/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.332632 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" path="/var/lib/kubelet/pods/12fa44d8-e695-4275-b09f-7bae13d6476d/volumes" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.363122 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-content/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.555303 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-content/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.559560 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/extract-utilities/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.832213 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7nrmp_5e831ea5-2e06-4bf6-a2d3-e3fb117168d0/registry-server/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.900406 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z5nmm_46f25710-1380-49bc-9a98-84a7b4f1f455/marketplace-operator/0.log" Mar 20 11:49:57 crc kubenswrapper[4846]: I0320 11:49:57.943072 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-utilities/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.131594 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-utilities/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.131631 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-content/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.189649 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-content/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.415252 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-content/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.449107 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/extract-utilities/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.520026 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nqpzh_30a01aab-f965-4b19-97b5-a3bccd0aaff3/registry-server/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.644042 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-utilities/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.858745 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-utilities/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.858762 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-content/0.log" Mar 20 11:49:58 crc kubenswrapper[4846]: I0320 11:49:58.892079 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-content/0.log" Mar 20 11:49:59 crc kubenswrapper[4846]: I0320 11:49:59.035043 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-utilities/0.log" Mar 20 11:49:59 crc kubenswrapper[4846]: I0320 11:49:59.091706 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/extract-content/0.log" Mar 20 11:49:59 crc kubenswrapper[4846]: I0320 11:49:59.569798 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bcsxf_70703261-65d2-40da-aa0d-948663787c75/registry-server/0.log" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.145668 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566790-dxxst"] Mar 20 11:50:00 crc kubenswrapper[4846]: E0320 11:50:00.146090 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="extract-utilities" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.146106 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="extract-utilities" Mar 20 11:50:00 crc kubenswrapper[4846]: E0320 11:50:00.146118 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="registry-server" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.146124 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="registry-server" Mar 20 11:50:00 crc kubenswrapper[4846]: E0320 11:50:00.146151 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="extract-content" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.146160 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="extract-content" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.146287 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12fa44d8-e695-4275-b09f-7bae13d6476d" containerName="registry-server" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.146853 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.151635 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.151635 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.154888 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.160257 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566790-dxxst"] Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.303194 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g86dr\" (UniqueName: \"kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr\") pod \"auto-csr-approver-29566790-dxxst\" (UID: \"2bdee2ae-a8c1-436b-94cb-4da15419e10f\") " pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.405041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g86dr\" (UniqueName: \"kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr\") pod \"auto-csr-approver-29566790-dxxst\" (UID: \"2bdee2ae-a8c1-436b-94cb-4da15419e10f\") " pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.432814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g86dr\" (UniqueName: \"kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr\") pod \"auto-csr-approver-29566790-dxxst\" (UID: \"2bdee2ae-a8c1-436b-94cb-4da15419e10f\") " pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.469532 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:00 crc kubenswrapper[4846]: I0320 11:50:00.933289 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566790-dxxst"] Mar 20 11:50:01 crc kubenswrapper[4846]: I0320 11:50:01.752360 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566790-dxxst" event={"ID":"2bdee2ae-a8c1-436b-94cb-4da15419e10f","Type":"ContainerStarted","Data":"ae50e9cbcf41837899839f5ae8ace7ea047317b9d9dad36e224ef32beda4e2e8"} Mar 20 11:50:02 crc kubenswrapper[4846]: I0320 11:50:02.765881 4846 generic.go:334] "Generic (PLEG): container finished" podID="2bdee2ae-a8c1-436b-94cb-4da15419e10f" containerID="6dfd844ab881c86a3d35654b8eef79e0a5fa6e7e7b2e70f0f50e9e459ea2144b" exitCode=0 Mar 20 11:50:02 crc kubenswrapper[4846]: I0320 11:50:02.766046 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566790-dxxst" event={"ID":"2bdee2ae-a8c1-436b-94cb-4da15419e10f","Type":"ContainerDied","Data":"6dfd844ab881c86a3d35654b8eef79e0a5fa6e7e7b2e70f0f50e9e459ea2144b"} Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.087035 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.270563 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g86dr\" (UniqueName: \"kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr\") pod \"2bdee2ae-a8c1-436b-94cb-4da15419e10f\" (UID: \"2bdee2ae-a8c1-436b-94cb-4da15419e10f\") " Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.280815 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr" (OuterVolumeSpecName: "kube-api-access-g86dr") pod "2bdee2ae-a8c1-436b-94cb-4da15419e10f" (UID: "2bdee2ae-a8c1-436b-94cb-4da15419e10f"). InnerVolumeSpecName "kube-api-access-g86dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.372597 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g86dr\" (UniqueName: \"kubernetes.io/projected/2bdee2ae-a8c1-436b-94cb-4da15419e10f-kube-api-access-g86dr\") on node \"crc\" DevicePath \"\"" Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.783712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566790-dxxst" event={"ID":"2bdee2ae-a8c1-436b-94cb-4da15419e10f","Type":"ContainerDied","Data":"ae50e9cbcf41837899839f5ae8ace7ea047317b9d9dad36e224ef32beda4e2e8"} Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.783778 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae50e9cbcf41837899839f5ae8ace7ea047317b9d9dad36e224ef32beda4e2e8" Mar 20 11:50:04 crc kubenswrapper[4846]: I0320 11:50:04.783775 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566790-dxxst" Mar 20 11:50:05 crc kubenswrapper[4846]: I0320 11:50:05.164211 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566784-5mbbf"] Mar 20 11:50:05 crc kubenswrapper[4846]: I0320 11:50:05.173728 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566784-5mbbf"] Mar 20 11:50:05 crc kubenswrapper[4846]: I0320 11:50:05.331570 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a755a71b-dc8c-4bcd-9144-60454c21d8a4" path="/var/lib/kubelet/pods/a755a71b-dc8c-4bcd-9144-60454c21d8a4/volumes" Mar 20 11:50:08 crc kubenswrapper[4846]: I0320 11:50:08.324075 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:50:08 crc kubenswrapper[4846]: E0320 11:50:08.324686 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:50:18 crc kubenswrapper[4846]: I0320 11:50:18.753955 4846 scope.go:117] "RemoveContainer" containerID="99b7218166dff07b7fb42f8f4d469029c709c5de20e38764d68a67e5bf62b50e" Mar 20 11:50:21 crc kubenswrapper[4846]: I0320 11:50:21.323254 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:50:21 crc kubenswrapper[4846]: E0320 11:50:21.325284 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:50:35 crc kubenswrapper[4846]: I0320 11:50:35.323735 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:50:35 crc kubenswrapper[4846]: E0320 11:50:35.324569 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:50:48 crc kubenswrapper[4846]: I0320 11:50:48.323529 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:50:48 crc kubenswrapper[4846]: E0320 11:50:48.324722 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:50:59 crc kubenswrapper[4846]: I0320 11:50:59.326256 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:50:59 crc kubenswrapper[4846]: E0320 11:50:59.327196 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:51:10 crc kubenswrapper[4846]: I0320 11:51:10.322579 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:51:10 crc kubenswrapper[4846]: E0320 11:51:10.323891 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:51:22 crc kubenswrapper[4846]: I0320 11:51:22.368763 4846 generic.go:334] "Generic (PLEG): container finished" podID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerID="da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31" exitCode=0 Mar 20 11:51:22 crc kubenswrapper[4846]: I0320 11:51:22.368865 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lhdz8/must-gather-gkdck" event={"ID":"fb3fbb06-e0a8-46a1-8f80-565f59acb451","Type":"ContainerDied","Data":"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31"} Mar 20 11:51:22 crc kubenswrapper[4846]: I0320 11:51:22.370148 4846 scope.go:117] "RemoveContainer" containerID="da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31" Mar 20 11:51:23 crc kubenswrapper[4846]: I0320 11:51:23.153789 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lhdz8_must-gather-gkdck_fb3fbb06-e0a8-46a1-8f80-565f59acb451/gather/0.log" Mar 20 11:51:24 crc kubenswrapper[4846]: I0320 11:51:24.322832 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:51:24 crc kubenswrapper[4846]: E0320 11:51:24.323274 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:51:30 crc kubenswrapper[4846]: I0320 11:51:30.746310 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lhdz8/must-gather-gkdck"] Mar 20 11:51:30 crc kubenswrapper[4846]: I0320 11:51:30.747417 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lhdz8/must-gather-gkdck" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="copy" containerID="cri-o://fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad" gracePeriod=2 Mar 20 11:51:30 crc kubenswrapper[4846]: I0320 11:51:30.755918 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lhdz8/must-gather-gkdck"] Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.174229 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lhdz8_must-gather-gkdck_fb3fbb06-e0a8-46a1-8f80-565f59acb451/copy/0.log" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.174961 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.270825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhk9v\" (UniqueName: \"kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v\") pod \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.270942 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output\") pod \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\" (UID: \"fb3fbb06-e0a8-46a1-8f80-565f59acb451\") " Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.277648 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v" (OuterVolumeSpecName: "kube-api-access-lhk9v") pod "fb3fbb06-e0a8-46a1-8f80-565f59acb451" (UID: "fb3fbb06-e0a8-46a1-8f80-565f59acb451"). InnerVolumeSpecName "kube-api-access-lhk9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.369446 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fb3fbb06-e0a8-46a1-8f80-565f59acb451" (UID: "fb3fbb06-e0a8-46a1-8f80-565f59acb451"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.372591 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhk9v\" (UniqueName: \"kubernetes.io/projected/fb3fbb06-e0a8-46a1-8f80-565f59acb451-kube-api-access-lhk9v\") on node \"crc\" DevicePath \"\"" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.372629 4846 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fb3fbb06-e0a8-46a1-8f80-565f59acb451-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.551398 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lhdz8_must-gather-gkdck_fb3fbb06-e0a8-46a1-8f80-565f59acb451/copy/0.log" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.552040 4846 generic.go:334] "Generic (PLEG): container finished" podID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerID="fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad" exitCode=143 Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.552121 4846 scope.go:117] "RemoveContainer" containerID="fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.552154 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lhdz8/must-gather-gkdck" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.572146 4846 scope.go:117] "RemoveContainer" containerID="da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.623146 4846 scope.go:117] "RemoveContainer" containerID="fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad" Mar 20 11:51:31 crc kubenswrapper[4846]: E0320 11:51:31.624741 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad\": container with ID starting with fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad not found: ID does not exist" containerID="fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.625003 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad"} err="failed to get container status \"fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad\": rpc error: code = NotFound desc = could not find container \"fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad\": container with ID starting with fb80bea27b33e19960c5d6fe4e336c7750e423d2de54e8419a7a0d332d7a98ad not found: ID does not exist" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.625028 4846 scope.go:117] "RemoveContainer" containerID="da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31" Mar 20 11:51:31 crc kubenswrapper[4846]: E0320 11:51:31.625628 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31\": container with ID starting with da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31 not found: ID does not exist" containerID="da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31" Mar 20 11:51:31 crc kubenswrapper[4846]: I0320 11:51:31.625683 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31"} err="failed to get container status \"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31\": rpc error: code = NotFound desc = could not find container \"da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31\": container with ID starting with da8ecddd44f7938a40ce523f9b28b3324e4394a911721873b0133b1f3d3e1a31 not found: ID does not exist" Mar 20 11:51:33 crc kubenswrapper[4846]: I0320 11:51:33.331735 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" path="/var/lib/kubelet/pods/fb3fbb06-e0a8-46a1-8f80-565f59acb451/volumes" Mar 20 11:51:36 crc kubenswrapper[4846]: I0320 11:51:36.322793 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:51:36 crc kubenswrapper[4846]: E0320 11:51:36.323515 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:51:51 crc kubenswrapper[4846]: I0320 11:51:51.322639 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:51:51 crc kubenswrapper[4846]: E0320 11:51:51.324550 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.177259 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566792-xjwfh"] Mar 20 11:52:00 crc kubenswrapper[4846]: E0320 11:52:00.178513 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bdee2ae-a8c1-436b-94cb-4da15419e10f" containerName="oc" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178533 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bdee2ae-a8c1-436b-94cb-4da15419e10f" containerName="oc" Mar 20 11:52:00 crc kubenswrapper[4846]: E0320 11:52:00.178551 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="gather" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="gather" Mar 20 11:52:00 crc kubenswrapper[4846]: E0320 11:52:00.178605 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="copy" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178619 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="copy" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178859 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bdee2ae-a8c1-436b-94cb-4da15419e10f" containerName="oc" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178888 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="copy" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.178939 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3fbb06-e0a8-46a1-8f80-565f59acb451" containerName="gather" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.179722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.184188 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.184523 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.187096 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566792-xjwfh"] Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.190292 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.236961 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqfrv\" (UniqueName: \"kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv\") pod \"auto-csr-approver-29566792-xjwfh\" (UID: \"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a\") " pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.338334 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqfrv\" (UniqueName: \"kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv\") pod \"auto-csr-approver-29566792-xjwfh\" (UID: \"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a\") " pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.362457 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqfrv\" (UniqueName: \"kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv\") pod \"auto-csr-approver-29566792-xjwfh\" (UID: \"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a\") " pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.511811 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.940409 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566792-xjwfh"] Mar 20 11:52:00 crc kubenswrapper[4846]: I0320 11:52:00.949557 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 11:52:01 crc kubenswrapper[4846]: I0320 11:52:01.792740 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" event={"ID":"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a","Type":"ContainerStarted","Data":"ffdd2daae68c773312200bb5d0b5fa1a8d0b226c074da4051c979b9232be54d3"} Mar 20 11:52:02 crc kubenswrapper[4846]: I0320 11:52:02.323075 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:52:02 crc kubenswrapper[4846]: E0320 11:52:02.323317 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:52:03 crc kubenswrapper[4846]: I0320 11:52:03.809156 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4ad931e-7ccc-463e-9f40-2ef1f43ff59a" containerID="f7dd7276137ba4abc28e8acb1f2cd5a1ab7005b9abfb536b8f1fbc250b5d9470" exitCode=0 Mar 20 11:52:03 crc kubenswrapper[4846]: I0320 11:52:03.809207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" event={"ID":"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a","Type":"ContainerDied","Data":"f7dd7276137ba4abc28e8acb1f2cd5a1ab7005b9abfb536b8f1fbc250b5d9470"} Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.127610 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.213231 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqfrv\" (UniqueName: \"kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv\") pod \"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a\" (UID: \"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a\") " Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.224267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv" (OuterVolumeSpecName: "kube-api-access-cqfrv") pod "f4ad931e-7ccc-463e-9f40-2ef1f43ff59a" (UID: "f4ad931e-7ccc-463e-9f40-2ef1f43ff59a"). InnerVolumeSpecName "kube-api-access-cqfrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.315246 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqfrv\" (UniqueName: \"kubernetes.io/projected/f4ad931e-7ccc-463e-9f40-2ef1f43ff59a-kube-api-access-cqfrv\") on node \"crc\" DevicePath \"\"" Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.825369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" event={"ID":"f4ad931e-7ccc-463e-9f40-2ef1f43ff59a","Type":"ContainerDied","Data":"ffdd2daae68c773312200bb5d0b5fa1a8d0b226c074da4051c979b9232be54d3"} Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.825733 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffdd2daae68c773312200bb5d0b5fa1a8d0b226c074da4051c979b9232be54d3" Mar 20 11:52:05 crc kubenswrapper[4846]: I0320 11:52:05.825446 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566792-xjwfh" Mar 20 11:52:06 crc kubenswrapper[4846]: I0320 11:52:06.192096 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566786-zjpgg"] Mar 20 11:52:06 crc kubenswrapper[4846]: I0320 11:52:06.196705 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566786-zjpgg"] Mar 20 11:52:07 crc kubenswrapper[4846]: I0320 11:52:07.339390 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db43ed65-2a0a-4ed3-b37b-dea97fb65cd2" path="/var/lib/kubelet/pods/db43ed65-2a0a-4ed3-b37b-dea97fb65cd2/volumes" Mar 20 11:52:13 crc kubenswrapper[4846]: I0320 11:52:13.328205 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:52:13 crc kubenswrapper[4846]: E0320 11:52:13.329286 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:52:18 crc kubenswrapper[4846]: I0320 11:52:18.863683 4846 scope.go:117] "RemoveContainer" containerID="b420ff8e526514cbd95a25ef0482ec0e4c55f12daf2cb2161f4e455dc83d32f1" Mar 20 11:52:25 crc kubenswrapper[4846]: I0320 11:52:25.322405 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:52:25 crc kubenswrapper[4846]: E0320 11:52:25.323373 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:52:36 crc kubenswrapper[4846]: I0320 11:52:36.322294 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:52:36 crc kubenswrapper[4846]: E0320 11:52:36.323439 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jds6p_openshift-machine-config-operator(aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b)\"" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" Mar 20 11:52:49 crc kubenswrapper[4846]: I0320 11:52:49.326609 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:52:50 crc kubenswrapper[4846]: I0320 11:52:50.150004 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"2a3c682ea7c9e49d5b29f705d43385a2b6feccc9504cfb487b97ac982d328bab"} Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.321286 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:53:59 crc kubenswrapper[4846]: E0320 11:53:59.322735 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ad931e-7ccc-463e-9f40-2ef1f43ff59a" containerName="oc" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.322754 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ad931e-7ccc-463e-9f40-2ef1f43ff59a" containerName="oc" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.322938 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ad931e-7ccc-463e-9f40-2ef1f43ff59a" containerName="oc" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.324157 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.335808 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.449586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwfrv\" (UniqueName: \"kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.449645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.449694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.550645 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.550780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwfrv\" (UniqueName: \"kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.550814 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.551129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.551192 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.572457 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwfrv\" (UniqueName: \"kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv\") pod \"redhat-marketplace-lnxbp\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:53:59 crc kubenswrapper[4846]: I0320 11:53:59.657553 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.112668 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.139206 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566794-99jxc"] Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.140443 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.142814 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.143007 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.143393 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.145593 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566794-99jxc"] Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.262106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrdsp\" (UniqueName: \"kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp\") pod \"auto-csr-approver-29566794-99jxc\" (UID: \"5496edc4-9df0-436e-9e36-0770566fa935\") " pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.363422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrdsp\" (UniqueName: \"kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp\") pod \"auto-csr-approver-29566794-99jxc\" (UID: \"5496edc4-9df0-436e-9e36-0770566fa935\") " pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.382654 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrdsp\" (UniqueName: \"kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp\") pod \"auto-csr-approver-29566794-99jxc\" (UID: \"5496edc4-9df0-436e-9e36-0770566fa935\") " pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.507571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.683684 4846 generic.go:334] "Generic (PLEG): container finished" podID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerID="7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a" exitCode=0 Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.683734 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerDied","Data":"7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a"} Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.683758 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerStarted","Data":"956977f94b0dbf4a2ce1c796aae0e2ca89ff696b19f2aef0bffd34da5d6363cd"} Mar 20 11:54:00 crc kubenswrapper[4846]: I0320 11:54:00.940771 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566794-99jxc"] Mar 20 11:54:00 crc kubenswrapper[4846]: W0320 11:54:00.953277 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5496edc4_9df0_436e_9e36_0770566fa935.slice/crio-08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c WatchSource:0}: Error finding container 08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c: Status 404 returned error can't find the container with id 08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c Mar 20 11:54:01 crc kubenswrapper[4846]: I0320 11:54:01.693562 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566794-99jxc" event={"ID":"5496edc4-9df0-436e-9e36-0770566fa935","Type":"ContainerStarted","Data":"08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c"} Mar 20 11:54:01 crc kubenswrapper[4846]: I0320 11:54:01.696225 4846 generic.go:334] "Generic (PLEG): container finished" podID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerID="9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240" exitCode=0 Mar 20 11:54:01 crc kubenswrapper[4846]: I0320 11:54:01.696350 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerDied","Data":"9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240"} Mar 20 11:54:02 crc kubenswrapper[4846]: I0320 11:54:02.708030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerStarted","Data":"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc"} Mar 20 11:54:02 crc kubenswrapper[4846]: I0320 11:54:02.712843 4846 generic.go:334] "Generic (PLEG): container finished" podID="5496edc4-9df0-436e-9e36-0770566fa935" containerID="cb233d6aaf49c2f999ee0170f37e41125b223748a171e235766f54d60232d4a7" exitCode=0 Mar 20 11:54:02 crc kubenswrapper[4846]: I0320 11:54:02.713219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566794-99jxc" event={"ID":"5496edc4-9df0-436e-9e36-0770566fa935","Type":"ContainerDied","Data":"cb233d6aaf49c2f999ee0170f37e41125b223748a171e235766f54d60232d4a7"} Mar 20 11:54:02 crc kubenswrapper[4846]: I0320 11:54:02.731970 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnxbp" podStartSLOduration=2.199273438 podStartE2EDuration="3.731947718s" podCreationTimestamp="2026-03-20 11:53:59 +0000 UTC" firstStartedPulling="2026-03-20 11:54:00.685393157 +0000 UTC m=+3412.051801393" lastFinishedPulling="2026-03-20 11:54:02.218067437 +0000 UTC m=+3413.584475673" observedRunningTime="2026-03-20 11:54:02.726084954 +0000 UTC m=+3414.092493200" watchObservedRunningTime="2026-03-20 11:54:02.731947718 +0000 UTC m=+3414.098355954" Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.098976 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.216651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrdsp\" (UniqueName: \"kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp\") pod \"5496edc4-9df0-436e-9e36-0770566fa935\" (UID: \"5496edc4-9df0-436e-9e36-0770566fa935\") " Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.222959 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp" (OuterVolumeSpecName: "kube-api-access-vrdsp") pod "5496edc4-9df0-436e-9e36-0770566fa935" (UID: "5496edc4-9df0-436e-9e36-0770566fa935"). InnerVolumeSpecName "kube-api-access-vrdsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.319709 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrdsp\" (UniqueName: \"kubernetes.io/projected/5496edc4-9df0-436e-9e36-0770566fa935-kube-api-access-vrdsp\") on node \"crc\" DevicePath \"\"" Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.730987 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566794-99jxc" event={"ID":"5496edc4-9df0-436e-9e36-0770566fa935","Type":"ContainerDied","Data":"08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c"} Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.731033 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08479c976c83fff7854d0d42edc808063e53948d0f2d77ed1697122bf4f4217c" Mar 20 11:54:04 crc kubenswrapper[4846]: I0320 11:54:04.731032 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566794-99jxc" Mar 20 11:54:05 crc kubenswrapper[4846]: I0320 11:54:05.166182 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566788-8m5k9"] Mar 20 11:54:05 crc kubenswrapper[4846]: I0320 11:54:05.172005 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566788-8m5k9"] Mar 20 11:54:05 crc kubenswrapper[4846]: I0320 11:54:05.332661 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="103af8a1-8d20-47da-91d8-168573c714b3" path="/var/lib/kubelet/pods/103af8a1-8d20-47da-91d8-168573c714b3/volumes" Mar 20 11:54:09 crc kubenswrapper[4846]: I0320 11:54:09.659153 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:09 crc kubenswrapper[4846]: I0320 11:54:09.660088 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:09 crc kubenswrapper[4846]: I0320 11:54:09.704325 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:09 crc kubenswrapper[4846]: I0320 11:54:09.815386 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:09 crc kubenswrapper[4846]: I0320 11:54:09.944944 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:54:11 crc kubenswrapper[4846]: I0320 11:54:11.781837 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnxbp" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="registry-server" containerID="cri-o://625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc" gracePeriod=2 Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.235867 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.432567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwfrv\" (UniqueName: \"kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv\") pod \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.432794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content\") pod \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.432856 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities\") pod \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\" (UID: \"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32\") " Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.434155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities" (OuterVolumeSpecName: "utilities") pod "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" (UID: "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.443047 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv" (OuterVolumeSpecName: "kube-api-access-xwfrv") pod "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" (UID: "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32"). InnerVolumeSpecName "kube-api-access-xwfrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.460181 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" (UID: "9c9f4a2f-6458-468b-b01e-19cf3cfe9f32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.534056 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.534109 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.534120 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwfrv\" (UniqueName: \"kubernetes.io/projected/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32-kube-api-access-xwfrv\") on node \"crc\" DevicePath \"\"" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.792025 4846 generic.go:334] "Generic (PLEG): container finished" podID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerID="625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc" exitCode=0 Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.792073 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerDied","Data":"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc"} Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.792101 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnxbp" event={"ID":"9c9f4a2f-6458-468b-b01e-19cf3cfe9f32","Type":"ContainerDied","Data":"956977f94b0dbf4a2ce1c796aae0e2ca89ff696b19f2aef0bffd34da5d6363cd"} Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.792102 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnxbp" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.792122 4846 scope.go:117] "RemoveContainer" containerID="625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.820217 4846 scope.go:117] "RemoveContainer" containerID="9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.829536 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.839085 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnxbp"] Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.840036 4846 scope.go:117] "RemoveContainer" containerID="7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.865795 4846 scope.go:117] "RemoveContainer" containerID="625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc" Mar 20 11:54:12 crc kubenswrapper[4846]: E0320 11:54:12.866331 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc\": container with ID starting with 625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc not found: ID does not exist" containerID="625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.866363 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc"} err="failed to get container status \"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc\": rpc error: code = NotFound desc = could not find container \"625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc\": container with ID starting with 625521f9c2d7cad3147718ae8e6fb33efdc1d29720b70b8429b0900cdacc67cc not found: ID does not exist" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.866385 4846 scope.go:117] "RemoveContainer" containerID="9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240" Mar 20 11:54:12 crc kubenswrapper[4846]: E0320 11:54:12.866711 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240\": container with ID starting with 9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240 not found: ID does not exist" containerID="9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.866761 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240"} err="failed to get container status \"9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240\": rpc error: code = NotFound desc = could not find container \"9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240\": container with ID starting with 9e3dc33ac95428a63ef91a23c97dd00b7c3c91d861f64f29025cf5865a6b5240 not found: ID does not exist" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.866797 4846 scope.go:117] "RemoveContainer" containerID="7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a" Mar 20 11:54:12 crc kubenswrapper[4846]: E0320 11:54:12.867163 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a\": container with ID starting with 7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a not found: ID does not exist" containerID="7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a" Mar 20 11:54:12 crc kubenswrapper[4846]: I0320 11:54:12.867198 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a"} err="failed to get container status \"7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a\": rpc error: code = NotFound desc = could not find container \"7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a\": container with ID starting with 7555cad8307fffc0d51e37a3c89a173b9ce9d68edc6e0304f8e3c41c2e1b244a not found: ID does not exist" Mar 20 11:54:13 crc kubenswrapper[4846]: I0320 11:54:13.333065 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" path="/var/lib/kubelet/pods/9c9f4a2f-6458-468b-b01e-19cf3cfe9f32/volumes" Mar 20 11:54:18 crc kubenswrapper[4846]: I0320 11:54:18.957719 4846 scope.go:117] "RemoveContainer" containerID="f5843a287dcfa07c01224499ee4c3faa841556b2db84f82bf3fa86ddb5d18cf5" Mar 20 11:55:09 crc kubenswrapper[4846]: I0320 11:55:09.678016 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:55:09 crc kubenswrapper[4846]: I0320 11:55:09.678737 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:55:39 crc kubenswrapper[4846]: I0320 11:55:39.678033 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:55:39 crc kubenswrapper[4846]: I0320 11:55:39.678798 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.143981 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566796-qml4z"] Mar 20 11:56:00 crc kubenswrapper[4846]: E0320 11:56:00.144979 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5496edc4-9df0-436e-9e36-0770566fa935" containerName="oc" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.144992 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5496edc4-9df0-436e-9e36-0770566fa935" containerName="oc" Mar 20 11:56:00 crc kubenswrapper[4846]: E0320 11:56:00.145004 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="extract-utilities" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145010 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="extract-utilities" Mar 20 11:56:00 crc kubenswrapper[4846]: E0320 11:56:00.145021 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="registry-server" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145028 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="registry-server" Mar 20 11:56:00 crc kubenswrapper[4846]: E0320 11:56:00.145039 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="extract-content" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145044 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="extract-content" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145164 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5496edc4-9df0-436e-9e36-0770566fa935" containerName="oc" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145186 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c9f4a2f-6458-468b-b01e-19cf3cfe9f32" containerName="registry-server" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.145615 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.148306 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2cgrx" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.148616 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.148764 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.163721 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566796-qml4z"] Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.302079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtcqf\" (UniqueName: \"kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf\") pod \"auto-csr-approver-29566796-qml4z\" (UID: \"5e068f89-89e0-4ca7-a475-3f0bc4e1886a\") " pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.403387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtcqf\" (UniqueName: \"kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf\") pod \"auto-csr-approver-29566796-qml4z\" (UID: \"5e068f89-89e0-4ca7-a475-3f0bc4e1886a\") " pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.426242 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtcqf\" (UniqueName: \"kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf\") pod \"auto-csr-approver-29566796-qml4z\" (UID: \"5e068f89-89e0-4ca7-a475-3f0bc4e1886a\") " pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.465449 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:00 crc kubenswrapper[4846]: I0320 11:56:00.872335 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566796-qml4z"] Mar 20 11:56:01 crc kubenswrapper[4846]: I0320 11:56:01.603392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566796-qml4z" event={"ID":"5e068f89-89e0-4ca7-a475-3f0bc4e1886a","Type":"ContainerStarted","Data":"c49d91decac1d03ebde035a9624e42d00bffeed164861817fe215e1d04857160"} Mar 20 11:56:02 crc kubenswrapper[4846]: I0320 11:56:02.612957 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e068f89-89e0-4ca7-a475-3f0bc4e1886a" containerID="0283d5a2522b223654d1c6bb292a71b863ba1ff6b3b9754ca8079aaef035485d" exitCode=0 Mar 20 11:56:02 crc kubenswrapper[4846]: I0320 11:56:02.613037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566796-qml4z" event={"ID":"5e068f89-89e0-4ca7-a475-3f0bc4e1886a","Type":"ContainerDied","Data":"0283d5a2522b223654d1c6bb292a71b863ba1ff6b3b9754ca8079aaef035485d"} Mar 20 11:56:03 crc kubenswrapper[4846]: I0320 11:56:03.894348 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.057595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtcqf\" (UniqueName: \"kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf\") pod \"5e068f89-89e0-4ca7-a475-3f0bc4e1886a\" (UID: \"5e068f89-89e0-4ca7-a475-3f0bc4e1886a\") " Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.063741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf" (OuterVolumeSpecName: "kube-api-access-rtcqf") pod "5e068f89-89e0-4ca7-a475-3f0bc4e1886a" (UID: "5e068f89-89e0-4ca7-a475-3f0bc4e1886a"). InnerVolumeSpecName "kube-api-access-rtcqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.159528 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtcqf\" (UniqueName: \"kubernetes.io/projected/5e068f89-89e0-4ca7-a475-3f0bc4e1886a-kube-api-access-rtcqf\") on node \"crc\" DevicePath \"\"" Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.630763 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566796-qml4z" event={"ID":"5e068f89-89e0-4ca7-a475-3f0bc4e1886a","Type":"ContainerDied","Data":"c49d91decac1d03ebde035a9624e42d00bffeed164861817fe215e1d04857160"} Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.630822 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c49d91decac1d03ebde035a9624e42d00bffeed164861817fe215e1d04857160" Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.630855 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566796-qml4z" Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.973736 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566790-dxxst"] Mar 20 11:56:04 crc kubenswrapper[4846]: I0320 11:56:04.978639 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566790-dxxst"] Mar 20 11:56:05 crc kubenswrapper[4846]: I0320 11:56:05.337123 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bdee2ae-a8c1-436b-94cb-4da15419e10f" path="/var/lib/kubelet/pods/2bdee2ae-a8c1-436b-94cb-4da15419e10f/volumes" Mar 20 11:56:09 crc kubenswrapper[4846]: I0320 11:56:09.677736 4846 patch_prober.go:28] interesting pod/machine-config-daemon-jds6p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 11:56:09 crc kubenswrapper[4846]: I0320 11:56:09.678222 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 11:56:09 crc kubenswrapper[4846]: I0320 11:56:09.678280 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" Mar 20 11:56:09 crc kubenswrapper[4846]: I0320 11:56:09.680972 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a3c682ea7c9e49d5b29f705d43385a2b6feccc9504cfb487b97ac982d328bab"} pod="openshift-machine-config-operator/machine-config-daemon-jds6p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 11:56:09 crc kubenswrapper[4846]: I0320 11:56:09.681098 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" podUID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerName="machine-config-daemon" containerID="cri-o://2a3c682ea7c9e49d5b29f705d43385a2b6feccc9504cfb487b97ac982d328bab" gracePeriod=600 Mar 20 11:56:10 crc kubenswrapper[4846]: I0320 11:56:10.682595 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b" containerID="2a3c682ea7c9e49d5b29f705d43385a2b6feccc9504cfb487b97ac982d328bab" exitCode=0 Mar 20 11:56:10 crc kubenswrapper[4846]: I0320 11:56:10.682667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerDied","Data":"2a3c682ea7c9e49d5b29f705d43385a2b6feccc9504cfb487b97ac982d328bab"} Mar 20 11:56:10 crc kubenswrapper[4846]: I0320 11:56:10.683645 4846 scope.go:117] "RemoveContainer" containerID="c19d5dc5b98bc2d1104a4004b091900e39239b1944dd4e171f7151082dd47d60" Mar 20 11:56:10 crc kubenswrapper[4846]: I0320 11:56:10.684565 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jds6p" event={"ID":"aa8a1ece-2e0b-4e8a-ba55-e719e1fc333b","Type":"ContainerStarted","Data":"45137fc9aeb0f137456ef2b4e0dfb3dd132a12aa035af0bb5c5ca6a15d2dd04c"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157232744024457 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157232745017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157223453016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157223453015464 5ustar corecore